Re: seg fault

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Mon, 8 Dec 2014, Philipp von Strobl-Albeg wrote:
> Hi,
> 
> after using the ceph-cluster for months without any problems - thank you for
> that great piece of software -, i recognize one osd crashed with following
> output.
> What are the recommondations - Just Upgrading or is this not a bug on 0.67 ?

You are running the original release of 0.67.  I recommend upgrading to 
the latest dumpling (0.67.10 or .11, I forget which) and seeing if the 
problem persists.

sage



> 
> 
>  -1> 2014-11-08 04:24:51.127924 7f0d92897700  5 --OSD::tracker-- reqid:
> client.9016.1:5037242, seq: 3484524, time: 2014-11-08 04:24:51.127924, event:
> waiting_for_osdmap, request: osd_op(client.9016.1:5037242
> rb.0.1798.6b8b4567.000000000076 [write 602112~4096] 2.c90060c7 snapc 7=[]
> e554) v4
>      0> 2014-11-08 04:24:51.141626 7f0d88ff9700 -1 *** Caught signal
> (Segmentation fault) **
>  in thread 7f0d88ff9700
> 
>  ceph version 0.67 (e3b7bc5bce8ab330ec1661381072368af3c218a0)
>  1: ceph-osd() [0x802577]
>  2: (()+0x113d0) [0x7f0db94d93d0]
>  3: (std::string::compare(std::string const&) const+0xc) [0x7f0db7e81c4c]
>  4: (PGLog::check()+0x90) [0x76b8d0]
>  5: (PGLog::write_log(ObjectStore::Transaction&, hobject_t const&)+0x245)
> [0x7672c5]
>  6: (PG::append_log(std::vector<pg_log_entry_t, std::allocator<pg_log_entry_t>
> >&, eversion_t, ObjectStore::Transaction&)+0x31d) [0x71f03d]
>  7: (ReplicatedPG::do_op(std::tr1::shared_ptr<OpRequest>)+0x36f3) [0x623e63]
>  8: (PG::do_request(std::tr1::shared_ptr<OpRequest>,
> ThreadPool::TPHandle&)+0x619) [0x710a19]
>  9: (OSD::dequeue_op(boost::intrusive_ptr<PG>,
> std::tr1::shared_ptr<OpRequest>, ThreadPool::TPHandle&)+0x330) [0x6663f0]
>  10: (OSD::OpWQ::_process(boost::intrusive_ptr<PG>,
> ThreadPool::TPHandle&)+0x4a0) [0x67cbc0]
>  11: (ThreadPool::WorkQueueVal<std::pair<boost::intrusive_ptr<PG>,
> std::tr1::shared_ptr<OpRequest> >, boost::intrusive_ptr<PG>
> >::_void_process(void*, ThreadPool::TPHandle&)+0x9c) [0x6b893c]
>  12: (ThreadPool::worker(ThreadPool::WorkThread*)+0x4e6) [0x8bb156]
>  13: (ThreadPool::WorkThread::entry()+0x10) [0x8bcf60]
>  14: (()+0x91a7) [0x7f0db94d11a7]
>  15: (clone()+0x6d) [0x7f0db76072cd]
>  NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to
> interpret this.
> 
> --- logging levels ---
>    0/ 5 none
>    0/ 1 lockdep
>    0/ 1 context
>    1/ 1 crush
>    1/ 5 mds
>    1/ 5 mds_balancer
>    1/ 5 mds_locker
>    1/ 5 mds_log
>    1/ 5 mds_log_expire
>    1/ 5 mds_migrator
>    0/ 1 buffer
>    0/ 1 timer
>    0/ 1 filer
>    0/ 1 striper
>    0/ 1 objecter
>    0/ 5 rados
>    0/ 5 rbd
>    0/ 5 journaler
>    0/ 5 objectcacher
>    0/ 5 client
>    0/ 5 osd
>    0/ 5 optracker
>    0/ 5 objclass
>    1/ 3 filestore
>    1/ 3 journal
>    0/ 5 ms
>    1/ 5 mon
>    0/10 monc
>    1/ 5 paxos
>    0/ 5 tp
>    1/ 5 auth
>    1/ 5 crypto
>    1/ 1 finisher
>    1/ 5 heartbeatmap
>    1/ 5 perfcounter
>    1/ 5 rgw
>    1/ 5 hadoop
>    1/ 5 javaclient
>  1/ 5 asok
>    1/ 1 throttle
>   -2/-2 (syslog threshold)
>   -1/-1 (stderr threshold)
>   max_recent     10000
>   max_new         1000
>   log_file /var/log/ceph/ceph-osd.2.log
> 
> -- 
> Philipp Strobl
> http://www.pilarkto.net
> 
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 
> 
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux