seg fault

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

after using the ceph-cluster for months without any problems - thank you for that great piece of software -, i recognize one osd crashed with following output.
What are the recommondations - Just Upgrading or is this not a bug on 0.67 ?


-1> 2014-11-08 04:24:51.127924 7f0d92897700 5 --OSD::tracker-- reqid: client.9016.1:5037242, seq: 3484524, time: 2014-11-08 04:24:51.127924, event: waiting_for_osdmap, request: osd_op(client.9016.1:5037242 rb.0.1798.6b8b4567.000000000076 [write 602112~4096] 2.c90060c7 snapc 7=[] e554) v4 0> 2014-11-08 04:24:51.141626 7f0d88ff9700 -1 *** Caught signal (Segmentation fault) **
 in thread 7f0d88ff9700

 ceph version 0.67 (e3b7bc5bce8ab330ec1661381072368af3c218a0)
 1: ceph-osd() [0x802577]
 2: (()+0x113d0) [0x7f0db94d93d0]
 3: (std::string::compare(std::string const&) const+0xc) [0x7f0db7e81c4c]
 4: (PGLog::check()+0x90) [0x76b8d0]
5: (PGLog::write_log(ObjectStore::Transaction&, hobject_t const&)+0x245) [0x7672c5] 6: (PG::append_log(std::vector<pg_log_entry_t, std::allocator<pg_log_entry_t> >&, eversion_t, ObjectStore::Transaction&)+0x31d) [0x71f03d] 7: (ReplicatedPG::do_op(std::tr1::shared_ptr<OpRequest>)+0x36f3) [0x623e63] 8: (PG::do_request(std::tr1::shared_ptr<OpRequest>, ThreadPool::TPHandle&)+0x619) [0x710a19] 9: (OSD::dequeue_op(boost::intrusive_ptr<PG>, std::tr1::shared_ptr<OpRequest>, ThreadPool::TPHandle&)+0x330) [0x6663f0] 10: (OSD::OpWQ::_process(boost::intrusive_ptr<PG>, ThreadPool::TPHandle&)+0x4a0) [0x67cbc0] 11: (ThreadPool::WorkQueueVal<std::pair<boost::intrusive_ptr<PG>, std::tr1::shared_ptr<OpRequest> >, boost::intrusive_ptr<PG> >::_void_process(void*, ThreadPool::TPHandle&)+0x9c) [0x6b893c]
 12: (ThreadPool::worker(ThreadPool::WorkThread*)+0x4e6) [0x8bb156]
 13: (ThreadPool::WorkThread::entry()+0x10) [0x8bcf60]
 14: (()+0x91a7) [0x7f0db94d11a7]
 15: (clone()+0x6d) [0x7f0db76072cd]
NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this.

--- logging levels ---
   0/ 5 none
   0/ 1 lockdep
   0/ 1 context
   1/ 1 crush
   1/ 5 mds
   1/ 5 mds_balancer
   1/ 5 mds_locker
   1/ 5 mds_log
   1/ 5 mds_log_expire
   1/ 5 mds_migrator
   0/ 1 buffer
   0/ 1 timer
   0/ 1 filer
   0/ 1 striper
   0/ 1 objecter
   0/ 5 rados
   0/ 5 rbd
   0/ 5 journaler
   0/ 5 objectcacher
   0/ 5 client
   0/ 5 osd
   0/ 5 optracker
   0/ 5 objclass
   1/ 3 filestore
   1/ 3 journal
   0/ 5 ms
   1/ 5 mon
   0/10 monc
   1/ 5 paxos
   0/ 5 tp
   1/ 5 auth
   1/ 5 crypto
   1/ 1 finisher
   1/ 5 heartbeatmap
   1/ 5 perfcounter
   1/ 5 rgw
   1/ 5 hadoop
   1/ 5 javaclient
 1/ 5 asok
   1/ 1 throttle
  -2/-2 (syslog threshold)
  -1/-1 (stderr threshold)
  max_recent     10000
  max_new         1000
  log_file /var/log/ceph/ceph-osd.2.log

--
Philipp Strobl
http://www.pilarkto.net

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux