At a guess, this is something that has long since been fixed in dumpling, you probably want to upgrade to the current dumpling point release. -Sam On Mon, Dec 8, 2014 at 2:40 PM, Philipp von Strobl-Albeg <philipp@xxxxxxxxxxxx> wrote: > Hi, > > after using the ceph-cluster for months without any problems - thank you for > that great piece of software -, i recognize one osd crashed with following > output. > What are the recommondations - Just Upgrading or is this not a bug on 0.67 ? > > > -1> 2014-11-08 04:24:51.127924 7f0d92897700 5 --OSD::tracker-- reqid: > client.9016.1:5037242, seq: 3484524, time: 2014-11-08 04:24:51.127924, > event: waiting_for_osdmap, request: osd_op(client.9016.1:5037242 > rb.0.1798.6b8b4567.000000000076 [write 602112~4096] 2.c90060c7 snapc 7=[] > e554) v4 > 0> 2014-11-08 04:24:51.141626 7f0d88ff9700 -1 *** Caught signal > (Segmentation fault) ** > in thread 7f0d88ff9700 > > ceph version 0.67 (e3b7bc5bce8ab330ec1661381072368af3c218a0) > 1: ceph-osd() [0x802577] > 2: (()+0x113d0) [0x7f0db94d93d0] > 3: (std::string::compare(std::string const&) const+0xc) [0x7f0db7e81c4c] > 4: (PGLog::check()+0x90) [0x76b8d0] > 5: (PGLog::write_log(ObjectStore::Transaction&, hobject_t const&)+0x245) > [0x7672c5] > 6: (PG::append_log(std::vector<pg_log_entry_t, > std::allocator<pg_log_entry_t> >&, eversion_t, > ObjectStore::Transaction&)+0x31d) [0x71f03d] > 7: (ReplicatedPG::do_op(std::tr1::shared_ptr<OpRequest>)+0x36f3) [0x623e63] > 8: (PG::do_request(std::tr1::shared_ptr<OpRequest>, > ThreadPool::TPHandle&)+0x619) [0x710a19] > 9: (OSD::dequeue_op(boost::intrusive_ptr<PG>, > std::tr1::shared_ptr<OpRequest>, ThreadPool::TPHandle&)+0x330) [0x6663f0] > 10: (OSD::OpWQ::_process(boost::intrusive_ptr<PG>, > ThreadPool::TPHandle&)+0x4a0) [0x67cbc0] > 11: (ThreadPool::WorkQueueVal<std::pair<boost::intrusive_ptr<PG>, > std::tr1::shared_ptr<OpRequest> >, boost::intrusive_ptr<PG> >>::_void_process(void*, ThreadPool::TPHandle&)+0x9c) [0x6b893c] > 12: (ThreadPool::worker(ThreadPool::WorkThread*)+0x4e6) [0x8bb156] > 13: (ThreadPool::WorkThread::entry()+0x10) [0x8bcf60] > 14: (()+0x91a7) [0x7f0db94d11a7] > 15: (clone()+0x6d) [0x7f0db76072cd] > NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to > interpret this. > > --- logging levels --- > 0/ 5 none > 0/ 1 lockdep > 0/ 1 context > 1/ 1 crush > 1/ 5 mds > 1/ 5 mds_balancer > 1/ 5 mds_locker > 1/ 5 mds_log > 1/ 5 mds_log_expire > 1/ 5 mds_migrator > 0/ 1 buffer > 0/ 1 timer > 0/ 1 filer > 0/ 1 striper > 0/ 1 objecter > 0/ 5 rados > 0/ 5 rbd > 0/ 5 journaler > 0/ 5 objectcacher > 0/ 5 client > 0/ 5 osd > 0/ 5 optracker > 0/ 5 objclass > 1/ 3 filestore > 1/ 3 journal > 0/ 5 ms > 1/ 5 mon > 0/10 monc > 1/ 5 paxos > 0/ 5 tp > 1/ 5 auth > 1/ 5 crypto > 1/ 1 finisher > 1/ 5 heartbeatmap > 1/ 5 perfcounter > 1/ 5 rgw > 1/ 5 hadoop > 1/ 5 javaclient > 1/ 5 asok > 1/ 1 throttle > -2/-2 (syslog threshold) > -1/-1 (stderr threshold) > max_recent 10000 > max_new 1000 > log_file /var/log/ceph/ceph-osd.2.log > > -- > Philipp Strobl > http://www.pilarkto.net > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com