Hi,
after using the ceph-cluster for months without any problems - thank you
for that great piece of software -, i recognize one osd crashed with
following output.
What are the recommondations - Just Upgrading or is this not a bug on 0.67 ?
-1> 2014-11-08 04:24:51.127924 7f0d92897700 5 --OSD::tracker-- reqid:
client.9016.1:5037242, seq: 3484524, time: 2014-11-08 04:24:51.127924,
event: waiting_for_osdmap, request: osd_op(client.9016.1:5037242
rb.0.1798.6b8b4567.000000000076 [write 602112~4096] 2.c90060c7 snapc
7=[] e554) v4
0> 2014-11-08 04:24:51.141626 7f0d88ff9700 -1 *** Caught signal
(Segmentation fault) **
in thread 7f0d88ff9700
ceph version 0.67 (e3b7bc5bce8ab330ec1661381072368af3c218a0)
1: ceph-osd() [0x802577]
2: (()+0x113d0) [0x7f0db94d93d0]
3: (std::string::compare(std::string const&) const+0xc) [0x7f0db7e81c4c]
4: (PGLog::check()+0x90) [0x76b8d0]
5: (PGLog::write_log(ObjectStore::Transaction&, hobject_t
const&)+0x245) [0x7672c5]
6: (PG::append_log(std::vector<pg_log_entry_t,
std::allocator<pg_log_entry_t> >&, eversion_t,
ObjectStore::Transaction&)+0x31d) [0x71f03d]
7: (ReplicatedPG::do_op(std::tr1::shared_ptr<OpRequest>)+0x36f3)
[0x623e63]
8: (PG::do_request(std::tr1::shared_ptr<OpRequest>,
ThreadPool::TPHandle&)+0x619) [0x710a19]
9: (OSD::dequeue_op(boost::intrusive_ptr<PG>,
std::tr1::shared_ptr<OpRequest>, ThreadPool::TPHandle&)+0x330) [0x6663f0]
10: (OSD::OpWQ::_process(boost::intrusive_ptr<PG>,
ThreadPool::TPHandle&)+0x4a0) [0x67cbc0]
11: (ThreadPool::WorkQueueVal<std::pair<boost::intrusive_ptr<PG>,
std::tr1::shared_ptr<OpRequest> >, boost::intrusive_ptr<PG>
>::_void_process(void*, ThreadPool::TPHandle&)+0x9c) [0x6b893c]
12: (ThreadPool::worker(ThreadPool::WorkThread*)+0x4e6) [0x8bb156]
13: (ThreadPool::WorkThread::entry()+0x10) [0x8bcf60]
14: (()+0x91a7) [0x7f0db94d11a7]
15: (clone()+0x6d) [0x7f0db76072cd]
NOTE: a copy of the executable, or `objdump -rdS <executable>` is
needed to interpret this.
--- logging levels ---
0/ 5 none
0/ 1 lockdep
0/ 1 context
1/ 1 crush
1/ 5 mds
1/ 5 mds_balancer
1/ 5 mds_locker
1/ 5 mds_log
1/ 5 mds_log_expire
1/ 5 mds_migrator
0/ 1 buffer
0/ 1 timer
0/ 1 filer
0/ 1 striper
0/ 1 objecter
0/ 5 rados
0/ 5 rbd
0/ 5 journaler
0/ 5 objectcacher
0/ 5 client
0/ 5 osd
0/ 5 optracker
0/ 5 objclass
1/ 3 filestore
1/ 3 journal
0/ 5 ms
1/ 5 mon
0/10 monc
1/ 5 paxos
0/ 5 tp
1/ 5 auth
1/ 5 crypto
1/ 1 finisher
1/ 5 heartbeatmap
1/ 5 perfcounter
1/ 5 rgw
1/ 5 hadoop
1/ 5 javaclient
1/ 5 asok
1/ 1 throttle
-2/-2 (syslog threshold)
-1/-1 (stderr threshold)
max_recent 10000
max_new 1000
log_file /var/log/ceph/ceph-osd.2.log
--
Philipp Strobl
http://www.pilarkto.net
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com