On 8/8/2013 12:30 PM, Suresh Sadhu wrote:
Thanks Mike,Please find the output of two commands
sadhu@ubuntu3:~$ ls /var/lib/ceph/osd/ceph-0
^^^ that is a problem. It appears that osd.0 didn't get deployed
properly. To see an example of what structure should be there, do:
ls /var/lib/ceph/osd/ceph-1
ceph-0 should be similar to the apparently working ceph-1 on your cluster.
It should look similar to:
#ls /var/lib/ceph/osd/ceph-0
ceph_fsid
current
fsid
keyring
magic
ready
store_version
whoami
- Mike
sadhu@ubuntu3:~$ cat /etc/ceph/ceph.conf
[global]
fsid = 593dac9e-ce55-4803-acb4-2d32b4e0d3be
mon_initial_members = ubuntu3
mon_host = 10.147.41.3
#auth_supported = cephx
auth cluster required = cephx
auth service required = cephx
auth client required = cephx
osd_journal_size = 1024
filestore_xattr_use_omap = true
-----Original Message-----
From: Mike Dawson [mailto:mike.dawson@xxxxxxxxxxxx]
Sent: 08 August 2013 18:50
To: Suresh Sadhu
Cc: ceph-users@xxxxxxxxxxxxxx
Subject: Re: how to recover the osd.
Looks like you didn't get osd.0 deployed properly. Can you show:
- ls /var/lib/ceph/osd/ceph-0
- cat /etc/ceph/ceph.conf
Thanks,
Mike Dawson
Co-Founder & Director of Cloud Architecture Cloudapt LLC
6330 East 75th Street, Suite 170
Indianapolis, IN 46250
On 8/8/2013 9:13 AM, Suresh Sadhu wrote:
HI,
My storage health cluster is warning state , one of the osd is in down
state and even if I try to start the osd it fail to start
sadhu@ubuntu3:~$ ceph osd stat
e22: 2 osds: 1 up, 1 in
sadhu@ubuntu3:~$ ls /var/lib/ceph/osd/
ceph-0 ceph-1
sadhu@ubuntu3:~$ ceph osd tree
# id weight type name up/down reweight
-1 0.14 root default
-2 0.14 host ubuntu3
0 0.06999 osd.0 down 0
1 0.06999 osd.1 up 1
sadhu@ubuntu3:~$ sudo /etc/init.d/ceph -a start 0
/etc/init.d/ceph: 0. not found (/etc/ceph/ceph.conf defines ,
/var/lib/ceph defines )
sadhu@ubuntu3:~$ sudo /etc/init.d/ceph -a start osd.0
/etc/init.d/ceph: osd.0 not found (/etc/ceph/ceph.conf defines ,
/var/lib/ceph defines )
Ceph health status in warning mode.
pg 4.10 is active+degraded, acting [1]
pg 3.17 is active+degraded, acting [1]
pg 5.16 is active+degraded, acting [1]
pg 4.17 is active+degraded, acting [1]
pg 3.10 is active+degraded, acting [1]
recovery 62/124 degraded (50.000%)
mds.ceph@ubuntu3 at 10.147.41.3:6803/2148 is laggy/unresponsi
regards
sadhu
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com