Hi, I launched a ceph (ceph version 0.80.5) lab on my laptop with 7 disk for osd. Yesterday all works fine, and i can create new pool and mount them. But after reboot, the ceph now working, more specificly the osd not start, belows are logs [root at cephnode1 ~]# ceph-disk activate-all === osd.5 === Mounting xfs on cephnode1:/var/lib/ceph/osd/ceph-5 failed: 'timeout 30 /usr/bin/ceph -c /etc/ceph/ceph.conf --name=osd.5 --keyring=/var/lib/ceph/osd/ceph-5/keyring osd crush create-or-move -- 5 0.02 host=cephnode1 root=default' ceph-disk: Error: ceph osd start failed: Command '['/sbin/service', 'ceph', 'start', 'osd.5']' returned non-zero exit status 1 === osd.7 === Mounting xfs on cephnode1:/var/lib/ceph/osd/ceph-7 failed: 'timeout 30 /usr/bin/ceph -c /etc/ceph/ceph.conf --name=osd.7 --keyring=/var/lib/ceph/osd/ceph-7/keyring osd crush create-or-move -- 7 0.02 host=cephnode1 root=default' ceph-disk: Error: ceph osd start failed: Command '['/sbin/service', 'ceph', 'start', 'osd.7']' returned non-zero exit status 1 === osd.3 === Mounting xfs on cephnode1:/var/lib/ceph/osd/ceph-3 failed: 'timeout 30 /usr/bin/ceph -c /etc/ceph/ceph.conf --name=osd.3 --keyring=/var/lib/ceph/osd/ceph-3/keyring osd crush create-or-move -- 3 0.02 host=cephnode1 root=default' ceph-disk: Error: ceph osd start failed: Command '['/sbin/service', 'ceph', 'start', 'osd.3']' returned non-zero exit status 1 === osd.4 === Mounting xfs on cephnode1:/var/lib/ceph/osd/ceph-4 failed: 'timeout 30 /usr/bin/ceph -c /etc/ceph/ceph.conf --name=osd.4 --keyring=/var/lib/ceph/osd/ceph-4/keyring osd crush create-or-move -- 4 0.02 host=cephnode1 root=default' ceph-disk: Error: ceph osd start failed: Command '['/sbin/service', 'ceph', 'start', 'osd.4']' returned non-zero exit status 1 === osd.1 === Mounting xfs on cephnode1:/var/lib/ceph/osd/ceph-1 failed: 'timeout 30 /usr/bin/ceph -c /etc/ceph/ceph.conf --name=osd.1 --keyring=/var/lib/ceph/osd/ceph-1/keyring osd crush create-or-move -- 1 0.02 host=cephnode1 root=default' ceph-disk: Error: ceph osd start failed: Command '['/sbin/service', 'ceph', 'start', 'osd.1']' returned non-zero exit status 1 === osd.2 === Mounting xfs on cephnode1:/var/lib/ceph/osd/ceph-2 failed: 'timeout 30 /usr/bin/ceph -c /etc/ceph/ceph.conf --name=osd.2 --keyring=/var/lib/ceph/osd/ceph-2/keyring osd crush create-or-move -- 2 0.02 host=cephnode1 root=default' ceph-disk: Error: ceph osd start failed: Command '['/sbin/service', 'ceph', 'start', 'osd.2']' returned non-zero exit status 1 === osd.6 === Mounting xfs on cephnode1:/var/lib/ceph/osd/ceph-6 failed: 'timeout 30 /usr/bin/ceph -c /etc/ceph/ceph.conf --name=osd.6 --keyring=/var/lib/ceph/osd/ceph-6/keyring osd crush create-or-move -- 6 0.02 host=cephnode1 root=default' ceph-disk: Error: ceph osd start failed: Command '['/sbin/service', 'ceph', 'start', 'osd.6']' returned non-zero exit status 1 === osd.0 === Mounting xfs on cephnode1:/var/lib/ceph/osd/ceph-0 failed: 'timeout 30 /usr/bin/ceph -c /etc/ceph/ceph.conf --name=osd.0 --keyring=/var/lib/ceph/osd/ceph-0/keyring osd crush create-or-move -- 0 0.02 host=cephnode1 root=default' ceph-disk: Error: ceph osd start failed: Command '['/sbin/service', 'ceph', 'start', 'osd.0']' returned non-zero exit status 1 ceph-disk: Error: One or more partitions failed to activate [root at cephnode1 ~]# ps -aef | grep ceph root 2021 1 0 21:02 ? 00:00:03 /usr/bin/ceph-mon -i cephnode1 --pid-file /var/run/ceph/mon.cephnode1.pid -c /etc/ceph/ceph.conf --cluster ceph root 2110 1 0 21:02 ? 00:00:03 /usr/bin/ceph-mds -i cephnode1 --pid-file /var/run/ceph/mds.cephnode1.pid -c /etc/ceph/ceph.conf --cluster ceph root 6965 2278 0 21:31 pts/1 00:00:00 grep ceph Do you have any ideas ? ? --- Thanks, Yitao(?? ?) jiangyt.github.io -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140811/d1c6473a/attachment.htm>