For future reference, that error was because the active MDS server was in replay. I can't tell why it didn't move on to active from what you posted, but I imagine it just got a little stuck since restarting made it work out. -Greg On Tuesday, June 5, 2012 at 1:05 PM, Martin Wilderoth wrote: > Hello Again, > > I restarted the mds on all servers and then it worked again > > /Regards Martin > > > Hello > > > > > Hi Martin, > > > > > > On 06/05/2012 08:07 PM, Martin Wilderoth wrote: > > > > Hello > > > > > > > > Is there a way to recover this error. > > > > > > > > mount -t ceph 10.0.6.10:/ /mnt -vv -o name=admin,secret=XXXXXXXXXXXXXXXXXXXXXXX > > > > [ 506.640433] libceph: loaded (mon/osd proto 15/24, osdmap 5/6 5/6) > > > > [ 506.650594] ceph: loaded (mds proto 32) > > > > [ 506.652353] libceph: client0 fsid a9d5f9e1-4bb9-4fab-b79b-ba4457631b01 > > > > [ 506.670876] Intel AES-NI instructions are not detected. > > > > [ 506.678861] libceph: mon0 10.0.6.10:6789 session established > > > > mount: 10.0.6.10:/: can't read superblock > > > > > > > > > > > > Could you share some more information? For example the output from: ceph -s > > > > 2012-06-05 20:25:05.307914 pg v1189604: 1152 pgs: 1152 active+clean; 191 GB data, 393 GB used, 973 GB / 1379 GB avail > > 012-06-05 20:25:05.315871 mds e60: 1/1/1 up {0=c=up:replay}, 2 up:standby > > 2012-06-05 20:25:05.315965 osd e1106: 8 osds: 8 up, 8 in > > 2012-06-05 20:25:05.316165 log 2012-06-05 20:24:50.425527 mon.0 10.0.6.10:6789/0 75 : [INF] mds.? 10.0.6.11:6800/22974 up:boot > > 2012-06-05 20:25:05.316371 mon e1: 3 mons at {a=10.0.6.10:6789/0,b=10.0.6.11:6789/0,c=10.0.6.12:6789/0} > > > > > > > > > > Did you change anything to the cluster since it worked? And what version > > > are you running? > > > > > > > > I have not done any changes installed at version 0.46 upgraded earlier and have been testing with > > ceph and ceph-fuse and backuppc. It was during the ceph-fuse it hanged. > > > > Current version > > ceph version 0.47.2 (commit:8bf9fde89bd6ebc4b0645b2fe02dadb1c17ad372) > > > > > > One of my mds logs has 24G of data. > > > > > > Is it still running? > > I have restarted mds.a and mds.b they seems to be running. But not everything. > > mds.a was stoped not sure mds.b but it has a big logfile. > > > > > > > > > > > > > I have some rbd devices that I would like to keep. > > > > > > RBD doesn't use the MDS nor the POSIX filesystem, so you will probably > > > be fine, but we need the output of "ceph -s" first. > > > > > > Does this work? > > > $ rbd ls > > > > > > this works I'm still using the rbd with no problem > > > $ rados -p rbd ls > > > > > > seems to work reports something simmilar to > > rb.0.2.00000000052e > > rb.0.0.0000000002f2 > > rb.0.7.000000000345 > > rb.0.7.000000000896 > > rb.0.0.000000000102 > > rb.0.9.000000000172 > > rb.0.1.000000000350 > > rb.0.4.000000000180 > > rb.0.4.00000000068b > > rb.0.5.00000000054c > > rb.0.2.0000000001e1 > > > > > Wido > > > > > > > > > > > /Regards Martin > -- > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > the body of a message to majordomo@xxxxxxxxxxxxxxx (mailto:majordomo@xxxxxxxxxxxxxxx) > More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html