Re: need help in a recovering ceph

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

On 11/26/2011 04:31 PM, Maxim Mikheev wrote:
Hi Everyone,

My ceph instance was too slow and I tried to benchmark it with bonnie++.
Benchmark finished with error message and Ceph died.
I cannot mount ceph on any computers:
root@s3-2core:~# mount -t ceph -o
name=admin,secretfile=/etc/ceph/mycluster.secret 192.168.2.11:/ /root/mnt/
mount error 5 = Input/output error

What does your "dmesg" show? Any additional information?


I have 250Gb data in ceph, I can generate this data again but it will
take around 2 weeks.

The question how can I reanimate ceph? I like the idea of ceph but
unfortunately I am not alone in the project an we are close to give up
on using it.
I just recently start to use it and don't know what kind of information
I should submit. Here are what I know:

root@s2-8core:~# ceph health
2011-11-26 10:20:32.565809 mon <- [health]
2011-11-26 10:20:32.566140 mon.0 -> 'HEALTH_OK' (0)

Although the cluster seems healthy, what does 'ceph -s' show?

Wido

---------------------------------------------------------------------------------------------------

# ceph osd dump -o -
2011-11-26 10:04:46.922947 mon <- [osd,dump]
2011-11-26 10:04:46.935616 mon.0 -> 'dumped osdmap epoch 465' (0)
epoch 465
fsid c09c2197-3976-3779-d7b1-26700db70b68
created 2011-11-04 12:43:26.390483
modifed 2011-11-25 22:21:17.275421
flags full

pool 0 'data' rep size 2 crush_ruleset 0 object_hash rjenkins pg_num 192
pgp_num 192 lpg_num 2 lpgp_num 2 last_change 5 owner 0
crash_replay_interval 60
removed_snaps [2~2]
pool 1 'metadata' rep size 2 crush_ruleset 1 object_hash rjenkins pg_num
192 pgp_num 192 lpg_num 2 lpgp_num 2 last_change 1 owner 0
pool 2 'rbd' rep size 2 crush_ruleset 2 object_hash rjenkins pg_num 192
pgp_num 192 lpg_num 2 lpgp_num 2 last_change 1 owner 0

max_osd 3
osd.0 up in weight 1 up_from 462 up_thru 462 down_at 303
last_clean_interval [292,298) lost_at 461 192.168.2.10:6800/4059
192.168.2.10:6801/4059 192.168.2.10:6802/4059
osd.1 up in weight 1 up_from 296 up_thru 462 down_at 295
last_clean_interval [281,294) lost_at 199 192.168.2.11:6801/3363
192.168.2.11:6805/3363 192.168.2.11:6806/3363
osd.2 up in weight 1 up_from 272 up_thru 462 down_at 268
last_clean_interval [257,267) 192.168.2.12:6800/1097
192.168.2.12:6801/1097 192.168.2.12:6802/1097


wrote 1081 byte payload to -
----------------------------------------------------------------------------------------------------

root@s2-8core:~# cat /etc/ceph/ceph.conf
[global]
auth supported = cephx
keyring = /etc/ceph/mycluster.keyring
max open files = 131072
log file = /var/log/ceph/$name.log
pid file = /var/run/ceph/$name.pid
[mon]
keyring = /etc/ceph/$name.keyring
mon data = /srv/mon.$id
debug ms = 1
[mon.a]
host = s2-8core
mon addr = 192.168.2.11:6789
[mds]
keyring = /etc/ceph/$name.keyring
[mds.a]
host = s2-8core
[osd]
keyring = /etc/ceph/$name.keyring
osd data = /srv/osd.$id
osd journal = /srv/osd.$id/journal
osd journal size = 1000 ; journal size, in megabytes
[osd.0]
host = s1-2core
[osd.1]
host = s2-8core
[osd.2]
host = s3-2core

Thank you in an advance,
Max
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html

--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux