Hi all, We have installed Ceph 0.78 on 3 hosts running SL6, each serving OSD and MON daemons, 2 of them running MDS in active/backup. Since today the MON and MDS daemons are crashing every time (after a short while). Rebooting the nodes didn't help. The cluster was running for 1-2 weeks without this problem. This is what's in the log files: 2014-04-01 12:48:09.442819 7fa39cb1d700 0 log [INF] : mdsmap e67: 1/1/1 up {0=ceph001.cubone.os=up:active(laggy or crashed)} 2014-04-01 12:48:09.442897 7fa39cb1d700 1 mon.ceph001 at 0(leader).paxos(paxos active c 504009..504758) is_readable now=2014-04-01 12:48:09.442898 lease_expire=2014-04-01 12:48:14.442854 has v0 lc 504758 2014-04-01 12:48:15.340419 7fa39d51e700 1 mon.ceph001 at 0(leader).paxos(paxos updating c 504009..504758) is_readable now=2014-04-01 12:48:15.340420 lease_expire=2014-04-01 12:48:14.442854 has v0 lc 504758 2014-04-01 12:48:15.340438 7fa39d51e700 1 mon.ceph001 at 0(leader).paxos(paxos updating c 504009..504758) is_readable now=2014-04-01 12:48:15.340439 lease_expire=2014-04-01 12:48:14.442854 has v0 lc 504758 2014-04-01 12:48:15.340443 7fa39d51e700 1 mon.ceph001 at 0(leader).paxos(paxos updating c 504009..504758) is_readable now=2014-04-01 12:48:15.340444 lease_expire=2014-04-01 12:48:14.442854 has v0 lc 504758 2014-04-01 12:48:22.260674 7fa39d51e700 1 mon.ceph001 at 0(leader).paxos(paxos active c 504009..504759) is_readable now=2014-04-01 12:48:22.260676 lease_expire=2014-04-01 12:48:27.260625 has v0 lc 504759 2014-04-01 12:48:22.260688 7fa39d51e700 1 mon.ceph001 at 0(leader).paxos(paxos active c 504009..504759) is_readable now=2014-04-01 12:48:22.260689 lease_expire=2014-04-01 12:48:27.260625 has v0 lc 504759 2014-04-01 12:48:22.260703 7fa39d51e700 1 mon.ceph001 at 0(leader).paxos(paxos active c 504009..504759) is_readable now=2014-04-01 12:48:22.260704 lease_expire=2014-04-01 12:48:27.260625 has v0 lc 504759 2014-04-01 12:48:22.292363 7fa39cb1d700 0 log [INF] : mon.ceph001 calling new monitor election 2014-04-01 12:48:22.292428 7fa39cb1d700 1 mon.ceph001 at 0(electing).elector(43) init, la2014-04-01 12:48:27.450606 7fa39d51e700 -1 *** Caught signal (Bus error) ** in thread 7fa39d51e700 ceph version 0.78 (f6c746c314d7b87b8419b6e584c94bfe4511dbd4) 1: /usr/bin/ceph-mon() [0x85e619] 2: /lib64/libpthread.so.0() [0x34c360f710] 3: (memcpy()+0x15b) [0x34c3289aab] 4: (()+0x3d4b2) [0x7fa3a0ae54b2] 5: (leveldb::log::Writer::EmitPhysicalRecord(leveldb::log::RecordType, char const*, unsigned long)+0x125) [0x7fa3a0acbe85] 6: (leveldb::log::Writer::AddRecord(leveldb::Slice const&)+0xe2) [0x7fa3a0acc092] 7: (leveldb::DBImpl::Write(leveldb::WriteOptions const&, leveldb::WriteBatch*)+0x38e) [0x7fa3a0ac0bde] 8: (LevelDBStore::submit_transaction_sync(std::tr1::shared_ptr<KeyValueDB::TransactionImpl>)+0x2b) [0x7c63eb] 9: (MonitorDBStore::apply_transaction(MonitorDBStore::Transaction const&)+0x183) [0x53c6b3] 10: (Paxos::begin(ceph::buffer::list&)+0x604) [0x5a2ca4] 11: (Paxos::propose_queued()+0x273) [0x5a3773] 12: (Paxos::propose_new_value(ceph::buffer::list&, Context*)+0x160) [0x5a3960] 13: (PaxosService::propose_pending()+0x386) [0x5ad946] 14: (Context::complete(int)+0x9) [0x580609] 15: (SafeTimer::timer_thread()+0x453) [0x768a33] 16: (SafeTimerThread::entry()+0xd) [0x76abed] 17: /lib64/libpthread.so.0() [0x34c36079d1] 18: (clone()+0x6d) [0x34c32e8b6d] NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this. --- begin dump of recent events --- -10000> 2014-04-01 12:37:41.891458 7fa39cb1d700 1 -- 10.141.8.180:6789/0 --> mon.2 10.141.8.182:6789/0 -- paxos(lease lc 504309 fc 503758 pn 0 opn 0) v3 -- ?+0 0x6246e00 -9999> 2014-04-01 12:37:41.891476 7fa39cb1d700 1 -- 10.141.8.180:6789/0 <== mon.0 10.141.8.180:6789/0 0 ==== log(1 entries) v1 ==== 0+0+0 (0 0 0) 0x29f4380 con 0x2050840 -9998> 2014-04-01 12:37:41.891487 7fa39cb1d700 1 mon.ceph001 at 0(leader).paxos(paxos active c 503758..504309) is_readable now=2014-04-01 12:37:41.891488 lease_expire=2014-04-01 12:37:46.891457 has v0 lc 504309 -9997> 2014-04-01 12:37:41.893979 7fa39cb1d700 1 -- 10.141.8.180:6789/0 <== mon.2 10.141.8.182:6789/0 672234191 ==== paxos(lease_ack lc 504309 fc 503758 pn 0 opn 0) v3 ==== 80+0+0 (2650854818 0 0) 0x6246e00 con 0x2051760 -9996> 2014-04-01 12:37:42.087358 7fa39cb1d700 1 -- 10.141.8.180:6789/0 <== mon.2 10.141.8.182:6789/0 672234192 ==== forward(pg_stats(1 pgs tid 13980 v 0) v1 caps allow profile osd tid 57315 con_features 4398046511103) to leader v2 ==== 916+0+0 (3952263404 0 0) 0x6247580 con 0x2051760 -9995> 2014-04-01 12:37:42.087395 7fa39cb1d700 1 mon.ceph001 at 0(leader).paxos(paxos active c 503758..504309) is_readable now=2014-04-01 12:37:42.087396 lease_expire=2014-04-01 12:37:46.891457 has v0 lc 504309 -9994> 2014-04-01 12:37:42.890767 7fa39d51e700 5 mon.ceph001 at 0(leader).paxos(paxos active c 503758..504309) queue_proposal bl 16145 bytes; ctx = 0x1fa6e90 -9993> 2014-04-01 12:37:42.891438 7fa39d51e700 1 -- 10.141.8.180:6789/0 --> mon.2 10.141.8.182:6789/0 -- paxos(begin lc 504309 fc 0 pn 1800 opn 0) v3 -- ?+0 0x369d000 -9992> 2014-04-01 12:37:42.891520 7fa39d51e700 5 mon.ceph001 at 0(leader).paxos(paxos updating c 503758..504309) queue_proposal bl 1025 bytes; ctx = 0x1fa0af0 -9991> 2014-04-01 12:37:42.891527 7fa39d51e700 5 mon.ceph001 at 0(leader).paxos(paxos updating c 503758..504309) propose_new_value not active; proposal queued -9990> 2014-04-01 12:37:42.892656 7fa39cb1d700 1 -- 10.141.8.180:6789/0 <== mon.2 10.141.8.182:6789/0 672234193 ==== paxos(accept lc 504309 fc 0 pn 1800 opn 0) v3 ==== 80+0+0 (2064056753 0 0) 0x369d000 con 0x2051760 -9989> 2014-04-01 12:37:42.893660 7fa39cb1d700 1 -- 10.141.8.180:6789/0 --> mon.2 10.141.8.182:6789/0 -- paxos(commit lc 504310 fc 0 pn 1800 opn 0) v3 -- ?+0 0x6247580 -9988> 2014-04-01 12:37:42.894019 7fa39cb1d700 1 -- 10.141.8.180:6789/0 --> 10.141.8.180:6789/0 -- log(last 1348) v1 -- ?+0 0x41ddcc0 con 0x2050840 -9987> 2014-04-01 12:37:42.894038 7fa39cb1d700 1 -- 10.141.8.180:6789/0 --> mon.2 10.141.8.182:6789/0 -- paxos(lease lc 504310 fc 503758 pn 0 opn 0) v3 -- ?+0 0x6246e00 2 of the 3 MONS are down, the MDSs are down, but the OSDs are still running.. Does someone knows what is happening here? Thanks! Kind regards, Kenneth