Re: OSD down

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Wow... all that verbose stuff is «just» a read error ?

But ok, I will continue on the ceph-users list then ;)


Le dimanche 10 février 2013 à 10:10 -0800, Gregory Farnum a écrit :
> The OSD daemon is getting back EIO when it tries to do a read. Sounds like your disk is going bad. 
> -Greg
> 
> PS: This question is a good fit for the new ceph-users list. :)
> 
> 
> On Sunday, February 10, 2013 at 9:45 AM, Olivier Bonvalet wrote:
> 
> > Hi,
> > 
> > I have an OSD which often stopped (ceph 0.56.2), with that in logs :
> > 
> > 446 stamp 2013-02-10 18:37:27.559777) v2 ==== 47+0+0 (4068038983 0 0)
> > 0x11e028c0 con 0x573d6e0
> > -3> 2013-02-10 18:37:27.561618 7f1c765d5700 1 --
> > 192.168.42.1:0/5824 <== osd.31 192.168.42.3:6811/23050 129 ====
> > osd_ping(ping_reply e13446 stamp 2013-02-10 18:37:27.559777) v2 ==== 47
> > +0+0 (4068038983 0 0) 0x73be380 con 0x573d420
> > -2> 2013-02-10 18:37:27.562674 7f1c765d5700 1 --
> > 192.168.42.1:0/5824 <== osd.1 192.168.42.2:6803/7458 129 ====
> > osd_ping(ping_reply e13446 stamp 2013-02-10 18:37:27.559777) v2 ==== 47
> > +0+0 (4068038983 0 0) 0x6bd8a80 con 0x573dc60
> > -1> 2013-02-10 18:37:28.217626 7f1c805e9700 5 osd.12 13444 tick
> > 0> 2013-02-10 18:37:28.552692 7f1c725cd700 -1 os/FileStore.cc (http://FileStore.cc): In
> > function 'virtual int FileStore::read(coll_t, const hobject_t&,
> > uint64_t, size_t, ceph::bufferlist&)' thread 7f1c725cd700 time
> > 2013-02-10 18:37:28.537715
> > os/FileStore.cc (http://FileStore.cc): 2732: FAILED assert(!m_filestore_fail_eio || got != -5)
> > 
> > ceph version ()
> > 1: (FileStore::read(coll_t, hobject_t const&, unsigned long, unsigned
> > long, ceph::buffer::list&)+0x462) [0x725f92]
> > 2: (PG::_scan_list(ScrubMap&, std::vector<hobject_t,
> > std::allocator<hobject_t> >&, bool)+0x371) [0x685da1]
> > 3: (PG::build_scrub_map_chunk(ScrubMap&, hobject_t, hobject_t,
> > bool)+0x29b) [0x6866bb]
> > 4: (PG::replica_scrub(MOSDRepScrub*)+0x8e9) [0x6952b9]
> > 5: (OSD::RepScrubWQ::_process(MOSDRepScrub*)+0xc2) [0x6410a2]
> > 6: (ThreadPool::worker(ThreadPool::WorkThread*)+0x879) [0x80f9e9]
> > 7: (ThreadPool::WorkThread::entry()+0x10) [0x8121f0]
> > 8: (()+0x68ca) [0x7f1c852f48ca]
> > 9: (clone()+0x6d) [0x7f1c83e23b6d]
> > NOTE: a copy of the executable, or `objdump -rdS <executable>` is
> > needed to interpret this.
> > 
> > --- logging levels ---
> > 0/ 5 none
> > 0/ 1 lockdep
> > 0/ 1 context
> > 1/ 1 crush
> > 1/ 5 mds
> > 1/ 5 mds_balancer
> > 1/ 5 mds_locker
> > 1/ 5 mds_log
> > 1/ 5 mds_log_expire
> > 1/ 5 mds_migrator
> > 0/ 1 buffer
> > 0/ 1 timer
> > 0/ 1 filer
> > 0/ 1 striper
> > 0/ 1 objecter
> > 0/ 5 rados
> > 0/ 5 rbd
> > 0/ 5 journaler
> > 0/ 5 objectcacher
> > 0/ 5 client
> > 0/ 5 osd
> > 0/ 5 optracker
> > 0/ 5 objclass
> > 1/ 3 filestore
> > 1/ 3 journal
> > 0/ 5 ms
> > 1/ 5 mon
> > 0/10 monc
> > 0/ 5 paxos
> > 0/ 5 tp
> > 1/ 5 auth
> > 1/ 5 crypto
> > 1/ 1 finisher
> > 1/ 5 heartbeatmap
> > 1/ 5 perfcounter
> > 1/ 5 rgw
> > 1/ 5 hadoop
> > 1/ 5 javaclient
> > 1/ 5 asok
> > 1/ 1 throttle
> > -2/-2 (syslog threshold)
> > -1/-1 (stderr threshold)
> > max_recent 100000
> > max_new 1000
> > log_file /var/log/ceph/osd.12.log
> > --- end dump of recent events ---
> > 2013-02-10 18:37:29.236649 7f1c725cd700 -1 *** Caught signal (Aborted)
> > **
> > in thread 7f1c725cd700
> > 
> > ceph version ()
> > 1: /usr/bin/ceph-osd() [0x7a0db9]
> > 2: (()+0xeff0) [0x7f1c852fcff0]
> > 3: (gsignal()+0x35) [0x7f1c83d861b5]
> > 4: (abort()+0x180) [0x7f1c83d88fc0]
> > 5: (__gnu_cxx::__verbose_terminate_handler()+0x115) [0x7f1c8461adc5]
> > 6: (()+0xcb166) [0x7f1c84619166]
> > 7: (()+0xcb193) [0x7f1c84619193]
> > 8: (()+0xcb28e) [0x7f1c8461928e]
> > 9: (ceph::__ceph_assert_fail(char const*, char const*, int, char
> > const*)+0x7c9) [0x8f3fc9]
> > 10: (FileStore::read(coll_t, hobject_t const&, unsigned long, unsigned
> > long, ceph::buffer::list&)+0x462) [0x725f92]
> > 11: (PG::_scan_list(ScrubMap&, std::vector<hobject_t,
> > std::allocator<hobject_t> >&, bool)+0x371) [0x685da1]
> > 12: (PG::build_scrub_map_chunk(ScrubMap&, hobject_t, hobject_t,
> > bool)+0x29b) [0x6866bb]
> > 13: (PG::replica_scrub(MOSDRepScrub*)+0x8e9) [0x6952b9]
> > 14: (OSD::RepScrubWQ::_process(MOSDRepScrub*)+0xc2) [0x6410a2]
> > 15: (ThreadPool::worker(ThreadPool::WorkThread*)+0x879) [0x80f9e9]
> > 16: (ThreadPool::WorkThread::entry()+0x10) [0x8121f0]
> > 17: (()+0x68ca) [0x7f1c852f48ca]
> > 18: (clone()+0x6d) [0x7f1c83e23b6d]
> > NOTE: a copy of the executable, or `objdump -rdS <executable>` is
> > needed to interpret this.
> > 
> > --- begin dump of recent events ---
> > -1> 2013-02-10 18:37:29.217778 7f1c805e9700 5 osd.12 13444 tick
> > 0> 2013-02-10 18:37:29.236649 7f1c725cd700 -1 *** Caught signal
> > (Aborted) **
> > in thread 7f1c725cd700
> > 
> > ceph version ()
> > 1: /usr/bin/ceph-osd() [0x7a0db9]
> > 2: (()+0xeff0) [0x7f1c852fcff0]
> > 3: (gsignal()+0x35) [0x7f1c83d861b5]
> > 4: (abort()+0x180) [0x7f1c83d88fc0]
> > 5: (__gnu_cxx::__verbose_terminate_handler()+0x115) [0x7f1c8461adc5]
> > 6: (()+0xcb166) [0x7f1c84619166]
> > 7: (()+0xcb193) [0x7f1c84619193]
> > 8: (()+0xcb28e) [0x7f1c8461928e]
> > 9: (ceph::__ceph_assert_fail(char const*, char const*, int, char
> > const*)+0x7c9) [0x8f3fc9]
> > 10: (FileStore::read(coll_t, hobject_t const&, unsigned long, unsigned
> > long, ceph::buffer::list&)+0x462) [0x725f92]
> > 11: (PG::_scan_list(ScrubMap&, std::vector<hobject_t,
> > std::allocator<hobject_t> >&, bool)+0x371) [0x685da1]
> > 12: (PG::build_scrub_map_chunk(ScrubMap&, hobject_t, hobject_t,
> > bool)+0x29b) [0x6866bb]
> > 13: (PG::replica_scrub(MOSDRepScrub*)+0x8e9) [0x6952b9]
> > 14: (OSD::RepScrubWQ::_process(MOSDRepScrub*)+0xc2) [0x6410a2]
> > 15: (ThreadPool::worker(ThreadPool::WorkThread*)+0x879) [0x80f9e9]
> > 16: (ThreadPool::WorkThread::entry()+0x10) [0x8121f0]
> > 17: (()+0x68ca) [0x7f1c852f48ca]
> > 18: (clone()+0x6d) [0x7f1c83e23b6d]
> > NOTE: a copy of the executable, or `objdump -rdS <executable>` is
> > needed to interpret this.
> > 
> > --- logging levels ---
> > 0/ 5 none
> > 0/ 1 lockdep
> > 0/ 1 context
> > 1/ 1 crush
> > 1/ 5 mds
> > 1/ 5 mds_balancer
> > 1/ 5 mds_locker
> > 1/ 5 mds_log
> > 1/ 5 mds_log_expire
> > 1/ 5 mds_migrator
> > 0/ 1 buffer
> > 0/ 1 timer
> > 0/ 1 filer
> > 0/ 1 striper
> > 0/ 1 objecter
> > 0/ 5 rados
> > 0/ 5 rbd
> > 0/ 5 journaler
> > 0/ 5 objectcacher
> > 0/ 5 client
> > 0/ 5 osd
> > 0/ 5 optracker
> > 0/ 5 objclass
> > 1/ 3 filestore
> > 1/ 3 journal
> > 0/ 5 ms
> > 1/ 5 mon
> > 0/10 monc
> > 0/ 5 paxos
> > 0/ 5 tp
> > 1/ 5 auth
> > 1/ 5 crypto
> > 1/ 1 finisher
> > 1/ 5 heartbeatmap
> > 1/ 5 perfcounter
> > 1/ 5 rgw
> > 1/ 5 hadoop
> > 1/ 5 javaclient
> > 1/ 5 asok
> > 1/ 1 throttle
> > -2/-2 (syslog threshold)
> > -1/-1 (stderr threshold)
> > max_recent 100000
> > max_new 1000
> > log_file /var/log/ceph/osd.12.log
> > --- end dump of recent events ---
> > 
> > 
> > 
> > What should I do ?
> > 
> > 
> > 
> > --
> > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> > the body of a message to majordomo@xxxxxxxxxxxxxxx (mailto:majordomo@xxxxxxxxxxxxxxx)
> > More majordomo info at http://vger.kernel.org/majordomo-info.html
> 
> 
> 
> 


--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux