Re: IO error: lock /var/lib/ceph/mon/ceph-ceph-node1/store.db/LOCK: Resource temporarily unavailable

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Joao,

On 12/05/2013 04:29 PM, Joao Eduardo Luis wrote:
On 12/05/2013 09:16 AM, Jan Kalcic wrote:
It seems ceph-mon does not exit with success, in fact:

ceph-node1 # sudo /usr/bin/ceph-mon -i ceph-node1 --pid-file
/var/run/ceph/mon.ceph-node1.pid -c /etc/ceph/ceph.conf -d
2013-12-05 10:06:27.429602 7fe06baf9780  0 ceph version 0.72.1
(4d923861868f6a15dcb33fef7f50f674997322de), process ceph-mon, pid 35662
IO error: lock /var/lib/ceph/mon/ceph-ceph-node1/store.db/LOCK: Resource
temporarily unavailable
IO error: lock /var/lib/ceph/mon/ceph-ceph-node1/store.db/LOCK: Resource
temporarily unavailable
2013-12-05 10:06:27.436295 7fe06baf9780 -1 failed to create new leveldb
store
ceph@ceph-node1:/var/run/ceph> sudo /usr/bin/ceph-mon -i ceph-node1
--pid-file /var/run/ceph/mon.ceph-node1.pid -c /etc/ceph/ceph.conf -d
IO error: lock /var/lib/ceph/mon/ceph-ceph-node1/store.db/LOCK: Resource
temporarily unavailable
IO error: lock /var/lib/ceph/mon/ceph-ceph-node1/store.db/LOCK: Resource
temporarily unavailable
2013-12-05 10:06:43.961756 7fd71d27c780  0 ceph version 0.72.1
(4d923861868f6a15dcb33fef7f50f674997322de), process ceph-mon, pid 36216
2013-12-05 10:06:43.964079 7fd71d27c780 -1 failed to create new leveldb
store

This is oftentime a symptom of another monitor running with the same id, and the monitor you're trying to start is unable to acquire the lock.


For example the following (3931, 3933) ??

# ps aux | grep ceph
root 3303 0.0 0.7 82184 3564 ? Ss 18:08 0:00 sshd: ceph [priv] ceph 3306 0.0 0.4 82312 2164 ? S 18:08 0:00 sshd: ceph@pts/0
ceph      3307  0.0  0.5  14004  2952 pts/0    Ss   18:08   0:00 -bash
root 3835 0.0 0.3 11648 1680 ? S 18:30 0:00 /bin/sh /usr/sbin/rcceph -c /etc/ceph/ceph.conf start mon.ceph-node1 root 3930 0.0 0.2 11512 1312 ? S 18:30 0:00 bash -c ulimit -n 32768; /usr/bin/ceph-mon -i ceph-node1 --pid-file /var/run/ceph/mon.ceph-node1.pid -c /etc/ceph/ceph.conf root 3931 0.0 1.0 76612 5328 ? S 18:30 0:00 /usr/bin/ceph-mon -i ceph-node1 --pid-file /var/run/ceph/mon.ceph-node1.pid -c /etc/ceph/ceph.conf root 3933 0.0 1.0 101688 5096 ? Sl 18:30 0:00 /usr/bin/ceph-mon -i ceph-node1 --pid-file /var/run/ceph/mon.ceph-node1.pid -c /etc/ceph/ceph.conf root 3941 0.0 1.7 69468 8796 ? S 18:30 0:00 python /usr/bin/ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.ceph-node1.asok mon_status
ceph      3964  0.0  0.1   4940   968 pts/0    R+   18:37   0:00 ps aux
ceph      3965  0.0  0.1   5712   820 pts/0    R+   18:37   0:00 grep ceph

but I am not doing tricky things, just following basic steps in the doc and running ceph-deploy as indicated.. what's wrong with ceph-deploy then??

Make sure /var/lib/ceph/mon/ceph-ceph-node1 exists and is a directory. Make sure you don't have any other monitor with id 'ceph-node1' running. If this error persists, set 'debug mon = 10' on ceph.conf, restart the monitor and send us the resulting log.

# find /var/lib/ceph/mon/ceph-ceph-node1
/var/lib/ceph/mon/ceph-ceph-node1
/var/lib/ceph/mon/ceph-ceph-node1/sysvinit
/var/lib/ceph/mon/ceph-ceph-node1/store.db
/var/lib/ceph/mon/ceph-ceph-node1/store.db/LOG
/var/lib/ceph/mon/ceph-ceph-node1/store.db/CURRENT
/var/lib/ceph/mon/ceph-ceph-node1/store.db/000006.log
/var/lib/ceph/mon/ceph-ceph-node1/store.db/000005.ldb
/var/lib/ceph/mon/ceph-ceph-node1/store.db/MANIFEST-000004
/var/lib/ceph/mon/ceph-ceph-node1/store.db/LOCK
/var/lib/ceph/mon/ceph-ceph-node1/done
/var/lib/ceph/mon/ceph-ceph-node1/keyring

# cat ceph.conf
[global]
filestore_xattr_use_omap = true
mon_host = 172.17.43.11
fsid = 18203301-f70b-40f1-b93e-fde4731f3bee
mon_initial_members = ceph-node1
auth_supported = cephx
osd_journal_size = 1024
debug mon = 10

...this is actually on ceph-admin but I have verified it is deployed afterward to ceph-node1 (by running --overwrite-conf)

# ceph-deploy --overwrite-conf mon create ceph-node1
[ceph_deploy.cli][INFO ] Invoked (1.3.3): /usr/bin/ceph-deploy --overwrite-conf mon create ceph-node1
[ceph_deploy.mon][DEBUG ] Deploying mon, cluster ceph hosts ceph-node1
[ceph_deploy.mon][DEBUG ] detecting platform for host ceph-node1 ...
[ceph-node1][DEBUG ] connected to host: ceph-node1
[ceph-node1][DEBUG ] detect platform information from remote host
[ceph-node1][DEBUG ] detect machine type
[ceph_deploy.mon][INFO ] distro info: SUSE Linux Enterprise Server 11 x86_64 [ceph-node1][DEBUG ] determining if provided host has same hostname in remote
[ceph-node1][DEBUG ] get remote short hostname
[ceph-node1][DEBUG ] deploying mon to ceph-node1
[ceph-node1][DEBUG ] get remote short hostname
[ceph-node1][DEBUG ] remote hostname: ceph-node1
[ceph-node1][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[ceph-node1][DEBUG ] create the mon path if it does not exist
[ceph-node1][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-ceph-node1/done
[ceph-node1][DEBUG ] create a done file to avoid re-doing the mon deployment
[ceph-node1][DEBUG ] create the init path if it does not exist
[ceph-node1][DEBUG ] locating the `service` executable...
[ceph-node1][INFO ] Running command: sudo rcceph -c /etc/ceph/ceph.conf start mon.ceph-node1
[ceph-node1][WARNIN] No data was received after 7 seconds, disconnecting...
[ceph-node1][INFO ] Running command: sudo ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.ceph-node1.asok mon_status

# cat /var/log/ceph/ceph-mon.ceph-node1.log
2013-12-05 18:30:36.473193 7f8f0ada3780 0 ceph version 0.72.1 (4d923861868f6a15dcb33fef7f50f674997322de), process ceph-mon, pid 3931

Not a lot of info in there though :( I see it is PID 3931, so 3933 might be the unwanted one?

Thanks,
Jan


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux