Hello, need help to solve segfault on all osd in my test cluster. Setup ceph from scratch. service ceph -a start ceph -w health HEALTH_OK monmap e1: 3 mons at {1=10.200.20.1:6789/0,2=10.200.20.2:6789/0,3=10.200.20.3:6789/0}, election epoch 6, quorum 0,1,2 1,2,3 osdmap e5: 4 osds: 4 up, 4 in pgmap v305: 960 pgs: 960 active+clean; 0 bytes data, 40147 MB used, 26667 GB / 26706 GB avail mdsmap e1: 0/0/1 up if i set replica size to 0 "i know this make no sense" ceph osd pool set rbd size 0 and then back to 2 ceph osd pool set rbd size 2 then i see that on all OSDs the process ceph-osd crash with segfault If i stop MONs daemons then i can start OSDs but if i start MONs back then die all OSDs again. How i cann repair this behavior ? My setup Nothing specials: Centos 6.3 Kernel: 3.8.3-1.el6.elrepo.x86_64 ceph-fuse-0.56.4-0.el6.x86_64 ceph-test-0.56.4-0.el6.x86_64 libcephfs1-0.56.4-0.el6.x86_64 ceph-0.56.4-0.el6.x86_64 ceph-release-1-0.el6.noarch cat /etc/ceph/ceph.conf [global] auth cluster required = none auth service required = none auth client required = none keyring = /etc/ceph/$name.keyring [mon] [mds] [osd] osd journal size = 10000 [mon.1] host = ceph-mon1 mon addr = 10.200.20.1:6789 [mon.2] host = ceph-mon2 mon addr = 10.200.20.2:6789 [mon.3] host = ceph-mon3 mon addr = 10.200.20.3:6789 [osd.0] host = ceph-osd1 [osd.1] host = ceph-osd2 [osd.2] host = ceph-osd3 [osd.3] host = ceph-osd4 [mds.a] host = ceph-mds1 [mds.b] host = ceph-mds2 [mds.c] host = ceph-mds3 Thanks much. -- AIXIT GmbH - Witalij Poljatchek (T) +49 69 203 4709-13 - (F) +49 69 203 470 979 wp@xxxxxxxxx - http://www.aixit.com AIXIT GmbH Strahlenbergerstr. 14 63067 Offenbach am Main (T) +49 69 203 470 913 Amtsgericht Offenbach, HRB 43953 Geschäftsführer: Friedhelm Heyer, Holger Grauer |
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com