Hi,
My ceph cluster include 5 OSDs. 3 osds are installed in the host 'strony-tc' and 2 are in the host 'strony-pc'. Recently, both of hosts were rebooted due to power cycles. After all of disks are mounted again, the ceph-osd are in the 'down' status. I tried cmd, "sudo start ceph-osd id=x', to start the OSDs. But they are not started well with the error below reported in the 'dmesg' output. Any suggestions about how to make the OSDs started well? Any comments are appreciated.
"
[6595400.895147] init: ceph-osd (ceph/1) main process ended, respawning
[6595400.969346] init: ceph-osd (ceph/1) main process (21990) terminated with status 1
[6595400.969352] init: ceph-osd (ceph/1) respawning too fast, stopped
"
:~$ ceph osd tree
ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY
-1 1.09477 root default
-2 0.61818 host strony-tc
0 0.20000 osd.0 down 0 1.00000
1 0.21819 osd.1 down 0 1.00000
4 0.20000 osd.4 up 1.00000 1.00000
-3 0.47659 host strony-pc
2 0.23830 osd.2 down 0 1.00000
3 0.23830 osd.3 down 0 1.00000
:~$ cat /etc/ceph/ceph.conf
[global]
fsid = 60638bfd-1eea-46d5-900d-36224475d8aa
mon_initial_members = strony-tc
mon_host = 10.132.141.122
auth_cluster_required = cephx
auth_service_required = cephx
auth_client_required = cephx
osd_pool_default_size = 2
Thanks,
Strony
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com