Hi, So the first error below is that /var/run/ceph isn't created when installing the ceph RPM(s). This is becasuse of line 440 in ceph.spec.in using the %ghost directive[1] for the file install. My reading of the behavior will mean that the file or directory in this case will be included in the manifest but not installed. I think this makes sense for specific log/state files but not the directory itself. I see two options 1) Remove the %ghost directive and allow RPM to install the directory. Potentially leaving orphaned pid/state files after the package is removed. 2) Or the directory needs to be created in the %post section. If it is created in the %post section and the %ghost directive stays then when the RPM is uninstalled any stale state pids/etc. will be automatically deleted and not left around. I personally would prefer the first option since that is the way most of the packages I have run into behave. This %ghost directive was new to me and I have written a large number of locally managed RPM packages. However there is still an error when this is fixed. This may be coming from the sysvinit 'service ceph start `hostname -s`' call at the end of mon_create. I can start the service by hand on all my monitors and then get further to where gatherkeys works and the ceph status seems to think the monitors is up. I am still looking to see how to track this exception but again seems very hard to with pushy. $ ../ceph-deploy/ceph-deploy mon create ceph00 Traceback (most recent call last): File "../ceph-deploy/ceph-deploy", line 8, in <module> load_entry_point('ceph-deploy==1.1', 'console_scripts', 'ceph-deploy')() File "/etc/ceph/ceph-deploy/ceph_deploy/cli.py", line 112, in main return args.func(args) File "/etc/ceph/ceph-deploy/ceph_deploy/mon.py", line 236, in mon mon_create(args) File "/etc/ceph/ceph-deploy/ceph_deploy/mon.py", line 140, in mon_create init=init, File "/etc/ceph/ceph-deploy/virtualenv/lib/python2.6/site-packages/pushy-0.5.1-py2.6.egg/pushy/protocol/proxy.py", line 255, in <lambda> (conn.operator(type_, self, args, kwargs)) File "/etc/ceph/ceph-deploy/virtualenv/lib/python2.6/site-packages/pushy-0.5.1-py2.6.egg/pushy/protocol/connection.py", line 66, in operator return self.send_request(type_, (object, args, kwargs)) File "/etc/ceph/ceph-deploy/virtualenv/lib/python2.6/site-packages/pushy-0.5.1-py2.6.egg/pushy/protocol/baseconnection.py", line 323, in send_request return self.__handle(m) File "/etc/ceph/ceph-deploy/virtualenv/lib/python2.6/site-packages/pushy-0.5.1-py2.6.egg/pushy/protocol/baseconnection.py", line 639, in __handle raise e ceph-mon: mon.noname-a 192.168.7.222:6789/0 is local, renaming to mon.ceph00 ceph-mon: set fsid to aee447eb-af2b-406c-b4f5-538a4cc84b3c ceph-mon: created monfs at /var/lib/ceph/mon/ceph-ceph00 for mon.ceph00 2013-06-18 14:54:24.265954 7f63ee659780 -1 asok(0x12d8000) AdminSocketConfigObs::init: failed: AdminSocket::bind_and_listen: failed to bind the UNIX domain socket to '/var/run/ceph/ceph-mon.ceph00.asok': (2) No such file or directory pushy.protocol.proxy.ExceptionProxy: [Errno 2] No such file or directory # ceph status health HEALTH_ERR 192 pgs stuck inactive; 192 pgs stuck unclean; no osds monmap e2: 3 mons at {ceph00=192.168.7.222:6789/0,ceph01=192.168.7.223:6789/0,ceph02=192.168.7.224:6789/0}, election epoch 6, quorum 0,1,2 ceph00,ceph01,ceph02 osdmap e1: 0 osds: 0 up, 0 in pgmap v2: 192 pgs: 192 creating; 0 bytes data, 0 KB used, 0 KB / 0 KB avail mdsmap e1: 0/0/1 up [1] - http://www.rpm.org/max-rpm-snapshot/s1-rpm-inside-files-list-directives.html -- --- Derek T. Yarnell University of Maryland Institute for Advanced Computer Studies _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com