Power Outage

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I can't really help with MDS.  Hopefully somebody else will chime in here.


On Tue, Aug 12, 2014 at 12:44 PM, hjcho616 <hjcho616 at yahoo.com> wrote:

> Craig,
>
> Thanks.  It turns out one of my memory stick went bad after that power
> outage.  While trying to fix the OSDs I ran in to many kernel crashes.
>  After removing that bad memory, I was able to fix them.  I did remove all
> OSD on that machine and rebuilt it as I didn't trust that data anymore. =P
>
> I was hoping MDS would come up after that.  But it didn't.  It shows this
> and kills itself.  Is this related to 0.82 MDS issue?
> 2014-08-12 14:35:11.250634 7ff794bd57c0  0 ceph version 0.80.5
> (38b73c67d375a2552d8ed67843c8a65c2c0feba6), process ceph-mds, pid 10244
> 2014-08-12 14:35:11.251092 7ff794bd57c0  1 -- 192.168.1.20:0/0 learned my
> addr 192.168.1.20:0/0
> 2014-08-12 14:35:11.251118 7ff794bd57c0  1 accepter.accepter.bind
> my_inst.addr is 192.168.1.20:6800/10244 need_addr=0
> 2014-08-12 14:35:11.259207 7ff794bd57c0  1 -- 192.168.1.20:6800/10244
> messenger.start
> 2014-08-12 14:35:11.259576 7ff794bd57c0 10 mds.-1.0 168 MDSCacheObject
> 2014-08-12 14:35:11.259625 7ff794bd57c0 10 mds.-1.0 2304        CInode
> 2014-08-12 14:35:11.259630 7ff794bd57c0 10 mds.-1.0 16   elist<>::item
> *7=112
> 2014-08-12 14:35:11.259635 7ff794bd57c0 10 mds.-1.0 480  inode_t
> 2014-08-12 14:35:11.259639 7ff794bd57c0 10 mds.-1.0 56    nest_info_t
> 2014-08-12 14:35:11.259644 7ff794bd57c0 10 mds.-1.0 32    frag_info_t
> 2014-08-12 14:35:11.259648 7ff794bd57c0 10 mds.-1.0 40   SimpleLock
> *5=200
> 2014-08-12 14:35:11.259652 7ff794bd57c0 10 mds.-1.0 48   ScatterLock
>  *3=144
> 2014-08-12 14:35:11.259656 7ff794bd57c0 10 mds.-1.0 488 CDentry
> 2014-08-12 14:35:11.259661 7ff794bd57c0 10 mds.-1.0 16   elist<>::item
> 2014-08-12 14:35:11.259669 7ff794bd57c0 10 mds.-1.0 40   SimpleLock
> 2014-08-12 14:35:11.259674 7ff794bd57c0 10 mds.-1.0 1016        CDir
> 2014-08-12 14:35:11.259678 7ff794bd57c0 10 mds.-1.0 16   elist<>::item
> *2=32
> 2014-08-12 14:35:11.259682 7ff794bd57c0 10 mds.-1.0 192  fnode_t
> 2014-08-12 14:35:11.259687 7ff794bd57c0 10 mds.-1.0 56    nest_info_t *2
> 2014-08-12 14:35:11.259691 7ff794bd57c0 10 mds.-1.0 32    frag_info_t *2
> 2014-08-12 14:35:11.259695 7ff794bd57c0 10 mds.-1.0 176 Capability
> 2014-08-12 14:35:11.259699 7ff794bd57c0 10 mds.-1.0 32   xlist<>::item
> *2=64
> 2014-08-12 14:35:11.259767 7ff794bd57c0  1 accepter.accepter.start
> 2014-08-12 14:35:11.260734 7ff794bd57c0  1 -- 192.168.1.20:6800/10244 -->
> 192.168.1.20:6789/0 -- auth(proto 0 31 bytes epoch 0) v1 -- ?+0 0x3684000
> con 0x36ac580
> 2014-08-12 14:35:11.261346 7ff794bcd700 10 mds.-1.0 MDS::ms_get_authorizer
> type=mon
> 2014-08-12 14:35:11.261696 7ff78fe4f700  5 mds.-1.0 ms_handle_connect on
> 192.168.1.20:6789/0
> 2014-08-12 14:35:11.262409 7ff78fe4f700  1 -- 192.168.1.20:6800/10244 <==
> mon.0 192.168.1.20:6789/0 1 ==== mon_map v1 ==== 194+0+0 (4155369063 0 0)
> 0x36d4000 con 0x36ac580
> 2014-08-12 14:35:11.262572 7ff78fe4f700  1 -- 192.168.1.20:6800/10244 <==
> mon.0 192.168.1.20:6789/0 2 ==== auth_reply(proto 2 0 (0) Success) v1
> ==== 33+0+0 (2093056952 0 0) 0x3691400 con 0x36ac580
> 2014-08-12 14:35:11.262925 7ff78fe4f700  1 -- 192.168.1.20:6800/10244 -->
> 192.168.1.20:6789/0 -- auth(proto 2 32 bytes epoch 0) v1 -- ?+0 0x3684240
> con 0x36ac580
> 2014-08-12 14:35:11.263643 7ff78fe4f700  1 -- 192.168.1.20:6800/10244 <==
> mon.0 192.168.1.20:6789/0 3 ==== auth_reply(proto 2 0 (0) Success) v1
> ==== 206+0+0 (1371651101 0 0) 0x3691800 con 0x36ac580
> 2014-08-12 14:35:11.263807 7ff78fe4f700  1 -- 192.168.1.20:6800/10244 -->
> 192.168.1.20:6789/0 -- auth(proto 2 165 bytes epoch 0) v1 -- ?+0
> 0x36846c0 con 0x36ac580
> 2014-08-12 14:35:11.264518 7ff78fe4f700  1 -- 192.168.1.20:6800/10244 <==
> mon.0 192.168.1.20:6789/0 4 ==== auth_reply(proto 2 0 (0) Success) v1
> ==== 580+0+0 (1904484134 0 0) 0x3691600 con 0x36ac580
> 2014-08-12 14:35:11.264662 7ff78fe4f700  1 -- 192.168.1.20:6800/10244 -->
> 192.168.1.20:6789/0 -- mon_subscribe({monmap=0+}) v2 -- ?+0 0x36b4380 con
> 0x36ac580
> 2014-08-12 14:35:11.264744 7ff78fe4f700  1 -- 192.168.1.20:6800/10244 -->
> 192.168.1.20:6789/0 -- auth(proto 2 2 bytes epoch 0) v1 -- ?+0 0x3684480
> con 0x36ac580
> 2014-08-12 14:35:11.265027 7ff78fe4f700  1 -- 192.168.1.20:6800/10244 <==
> mon.0 192.168.1.20:6789/0 5 ==== mon_map v1 ==== 194+0+0 (4155369063 0 0)
> 0x36d43c0 con 0x36ac580
> 2014-08-12 14:35:11.265203 7ff78fe4f700  1 -- 192.168.1.20:6800/10244 <==
> mon.0 192.168.1.20:6789/0 6 ==== mon_subscribe_ack(300s) v1 ==== 20+0+0
> (2253672535 0 0) 0x36b4540 con 0x36ac580
> 2014-08-12 14:35:11.265251 7ff78fe4f700  1 -- 192.168.1.20:6800/10244 <==
> mon.0 192.168.1.20:6789/0 7 ==== auth_reply(proto 2 0 (0) Success) v1
> ==== 194+0+0 (1999696020 0 0) 0x3691a00 con 0x36ac580
> 2014-08-12 14:35:11.265506 7ff794bd57c0  1 -- 192.168.1.20:6800/10244 -->
> 192.168.1.20:6789/0 -- mon_subscribe({monmap=2+,osdmap=0}) v2 -- ?+0
> 0x36b41c0 con 0x36ac580
> 2014-08-12 14:35:11.265580 7ff794bd57c0  1 -- 192.168.1.20:6800/10244 -->
> 192.168.1.20:6789/0 -- mon_subscribe({mdsmap=0+,monmap=2+,osdmap=0}) v2
> -- ?+0 0x36b4a80 con 0x36ac580
> 2014-08-12 14:35:11.266159 7ff78fe4f700  1 -- 192.168.1.20:6800/10244 <==
> mon.0 192.168.1.20:6789/0 8 ==== osd_map(9687..9687 src has 9090..9687)
> v3 ==== 6983+0+0 (1578463925 0 0) 0x3684b40 con 0x36ac580
> 2014-08-12 14:35:11.266453 7ff78fe4f700  1 -- 192.168.1.20:6800/10244 <==
> mon.0 192.168.1.20:6789/0 9 ==== mon_subscribe_ack(300s) v1 ==== 20+0+0
> (2253672535 0 0) 0x36b41c0 con 0x36ac580
> 2014-08-12 14:35:11.266491 7ff78fe4f700  1 -- 192.168.1.20:6800/10244 <==
> mon.0 192.168.1.20:6789/0 10 ==== mdsmap(e 7182) v1 ==== 653+0+0
> (374906493 0 0) 0x3691800 con 0x36ac580
> 2014-08-12 14:35:11.266518 7ff794bd57c0 10 mds.-1.0 beacon_send up:boot
> seq 1 (currently up:boot)
> 2014-08-12 14:35:11.266585 7ff794bd57c0  1 -- 192.168.1.20:6800/10244 -->
> 192.168.1.20:6789/0 -- mdsbeacon(12799/MDS1.1 up:boot seq 1 v0) v2 -- ?+0
> 0x36bc2c0 con 0x36ac580
> 2014-08-12 14:35:11.266626 7ff794bd57c0 10 mds.-1.0 create_logger
> 2014-08-12 14:35:11.266677 7ff78fe4f700  5 mds.-1.0 handle_mds_map epoch
> 7182 from mon.0
> 2014-08-12 14:35:11.266779 7ff78fe4f700 10 mds.-1.0      my compat
> compat={},rocompat={},incompat={1=base v0.20,2=client writeable
> ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds
> uses versioned encoding,6=dirfrag is stored in omap,7=mds uses inline data}
> 2014-08-12 14:35:11.266793 7ff78fe4f700 10 mds.-1.0  mdsmap compat
> compat={},rocompat={},incompat={1=base v0.20,2=client writeable
> ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds
> uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table}
> 2014-08-12 14:35:11.266803 7ff78fe4f700  0 mds.-1.0 handle_mds_map mdsmap
> compatset compat={},rocompat={},incompat={1=base v0.20,2=client writeable
> ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds
> uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table} not
> writeable with daemon features compat={},rocompat={},incompat={1=base
> v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode
> in separate object,5=mds uses versioned encoding,6=dirfrag is stored in
> omap,7=mds uses inline data}, killing myself
> 2014-08-12 14:35:11.266821 7ff78fe4f700  1 mds.-1.0 suicide.  wanted
> down:dne, now up:boot
> 2014-08-12 14:35:11.267081 7ff78fe4f700  1 -- 192.168.1.20:6800/10244
> mark_down 0x36ac580 -- 0x36c8500
> 2014-08-12 14:35:11.267204 7ff78fe4f700  1 -- 192.168.1.20:6800/10244
> mark_down_all
> 2014-08-12 14:35:11.267612 7ff794bd57c0  1 -- 192.168.1.20:6800/10244
> shutdown complete.
>
> Regards,
> Hong
>
>
>
>   On Tuesday, July 22, 2014 4:03 PM, Craig Lewis <
> clewis at centraldesktop.com> wrote:
>
>
>  The osd lost is useful, but not strictly required.  It accelerates the
> recovery once things are stable.  It tells Ceph to give up trying to
> recovery data off those disks.  Without it, Ceph will still check, then
> give up when it can't find it.
>
>
> I was having problems with the suicide timeout at one point.  Basically,
> the OSDs fail and restart so many times that they can't apply all of the
> map changes before they hit the timeout.  Sage gave me some suggestions.
>  Give this a try:
> https://www.mail-archive.com/ceph-devel at vger.kernel.org/msg18862.html
>
> That process solved suicide timeouts, with one caveat.  When I followed
> it, I filled up /var/log/ceph/ and the recovery failed.  I had to manually
> run each OSD in debugging mode until it completed the map update.  Aside
> from that, I followed your procedure.
>
> I had to run that procedure on all OSDs.  I did all OSDs on a node at the
> same time.
>
>
>
>
>
>
> On Mon, Jul 21, 2014 at 11:45 PM, hjcho616 <hjcho616 at yahoo.com> wrote:
>
> Craig,
>
> osd.2 was down and out.  lost wasn't working.. so skipped it. =P
>  Formatted the drive XFS and got mostly working but couldn't figure out how
> to get the journal to point at my SSD, and init script wasn't able to find
> the osd.2 for some reason.  So just used ceph-deploy.  It created new osd.6
> on the disks that were used for osd.2.  I removed norecover and nobackfill
> and let the system rebuild.  It seemed like it was doing well until it hit
> that suicide timeout. What should I do in this case?
>
>    -20> 2014-07-22 01:01:26.087707 7f3a90012700 10 monclient:
> _check_auth_rotating have uptodate secrets (they expire after 2014-07-22
> 01:00:56.087703)
>    -19> 2014-07-22 01:01:26.087743 7f3a90012700 10 monclient: renew subs?
> (now: 2014-07-22 01:01:26.087742; renew after: 2014-07-22 01:01:16.084357)
> -- yes
>    -18> 2014-07-22 01:01:26.087775 7f3a90012700 10 monclient: renew_subs
>    -17> 2014-07-22 01:01:26.087793 7f3a90012700 10 monclient:
> _send_mon_message to mon.MDS1 at 192.168.1.20:6789/0
>    -16> 2014-07-22 01:01:26.087822 7f3a90012700  1 --
> 192.168.1.30:6800/6297 --> 192.168.1.20:6789/0 --
> mon_subscribe({monmap=2+,osd_pg_creates=0}) v2 -- ?+0 0x1442c000 con
> 0xf73e2c0
>    -15> 2014-07-22 01:01:27.916972 7f3a8c80b700  5 osd.6 3252 heartbeat:
> osd_stat(66173 MB used, 1797 GB avail, 1862 GB total, peers [3,4,5]/[] op
> hist [])
>    -14> 2014-07-22 01:01:27.917061 7f3a8c80b700  1 -- 192.168.2.30:0/6297
> --> 192.168.2.31:6803/13623 -- osd_ping(ping e3252 stamp 2014-07-22
> 01:01:27.917024) v2 -- ?+0 0x140201c0 con 0xfa68160
>    -13> 2014-07-22 01:01:27.917131 7f3a8c80b700  1 -- 192.168.2.30:0/6297
> --> 192.168.1.31:6804/13623 -- osd_ping(ping e3252 stamp 2014-07-22
> 01:01:27.917024) v2 -- ?+0 0x17d0b500 con 0xfa68000
>    -12> 2014-07-22 01:01:27.917180 7f3a8c80b700  1 -- 192.168.2.30:0/6297
> --> 192.168.2.31:6805/13991 -- osd_ping(ping e3252 stamp 2014-07-22
> 01:01:27.917024) v2 -- ?+0 0xfa8fdc0 con 0x19208c60
>    -11> 2014-07-22 01:01:27.917229 7f3a8c80b700  1 -- 192.168.2.30:0/6297
> --> 192.168.1.31:6807/13991 -- osd_ping(ping e3252 stamp 2014-07-22
> 01:01:27.917024) v2 -- ?+0 0xffcee00 con 0x205c000
>    -10> 2014-07-22 01:01:27.917276 7f3a8c80b700  1 -- 192.168.2.30:0/6297
> --> 192.168.2.31:6801/13249 -- osd_ping(ping e3252 stamp 2014-07-22
> 01:01:27.917024) v2 -- ?+0 0x224fdc0 con 0xf9f8dc0
>     -9> 2014-07-22 01:01:27.917325 7f3a8c80b700  1 -- 192.168.2.30:0/6297
> --> 192.168.1.31:6801/13249 -- osd_ping(ping e3252 stamp 2014-07-22
> 01:01:27.917024) v2 -- ?+0 0xf8ce000 con 0x19208840
>     -8> 2014-07-22 01:01:27.918723 7f3a9581d700  1 -- 192.168.2.30:0/6297
> <== osd.3 192.168.1.31:6804/13623 28 ==== osd_ping(ping_reply e3252 stamp
> 2014-07-22 01:01:27.917024) v2 ==== 47+0+0 (2181285829 0 0) 0xffcf500 con
> 0xfa68000
>     -7> 2014-07-22 01:01:27.918830 7f3a9581d700  1 -- 192.168.2.30:0/6297
> <== osd.5 192.168.1.31:6801/13249 28 ==== osd_ping(ping_reply e3252 stamp
> 2014-07-22 01:01:27.917024) v2 ==== 47+0+0 (2181285829 0 0) 0x11a5e700 con
> 0x19208840
>     -6> 2014-07-22 01:01:27.919218 7f3a9581d700  1 -- 192.168.2.30:0/6297
> <== osd.5 192.168.2.31:6801/13249 28 ==== osd_ping(ping_reply e3252 stamp
> 2014-07-22 01:01:27.917024) v2 ==== 47+0+0 (2181285829 0 0) 0xfa8fa40 con
> 0xf9f8dc0
>     -5> 2014-07-22 01:01:27.919396 7f3a9581d700  1 -- 192.168.2.30:0/6297
> <== osd.3 192.168.2.31:6803/13623 28 ==== osd_ping(ping_reply e3252 stamp
> 2014-07-22 01:01:27.917024) v2 ==== 47+0+0 (2181285829 0 0) 0x1dd8bc00 con
> 0xfa68160
>     -4> 2014-07-22 01:01:27.919521 7f3a9581d700  1 -- 192.168.2.30:0/6297
> <== osd.4 192.168.2.31:6805/13991 28 ==== osd_ping(ping_reply e3252 stamp
> 2014-07-22 01:01:27.917024) v2 ==== 47+0+0 (2181285829 0 0) 0x14021a40 con
> 0x19208c60
>     -3> 2014-07-22 01:01:27.919606 7f3a9581d700  1 -- 192.168.2.30:0/6297
> <== osd.4 192.168.1.31:6807/13991 28 ==== osd_ping(ping_reply e3252 stamp
> 2014-07-22 01:01:27.917024) v2 ==== 47+0+0 (2181285829 0 0) 0x10fdd6c0 con
> 0x205c000
>     -2> 2014-07-22 01:01:29.976382 7f3aa5c22700  1 heartbeat_map
> is_healthy 'FileStore::op_tp thread 0x7f3a9d0b0700' had timed out after 60
>     -1> 2014-07-22 01:01:29.976416 7f3aa5c22700  1 heartbeat_map
> is_healthy 'FileStore::op_tp thread 0x7f3a9d0b0700' had suicide timed out
> after 180
>      0> 2014-07-22 01:01:29.985984 7f3aa5c22700 -1 common/HeartbeatMap.cc:
> In function 'bool ceph::HeartbeatMap::_check(ceph::heartbeat_handle_d*,
> const char*, time_t)' thread 7f3aa5c22700 time 2014-07-22 01:01:29.976450
> common/HeartbeatMap.cc: 79: FAILED assert(0 == "hit suicide timeout")
>
>  ceph version 0.82 (14085f42ddd0fef4e7e1dc99402d07a8df82c04e)
>  1: (ceph::HeartbeatMap::_check(ceph::heartbeat_handle_d*, char const*,
> long)+0x2eb) [0xad2cbb]
>  2: (ceph::HeartbeatMap::is_healthy()+0xb6) [0xad34c6]
>  3: (ceph::HeartbeatMap::check_touch_file()+0x28) [0xad3aa8]
>  4: (CephContextServiceThread::entry()+0x13f) [0xb9911f]
>  5: (()+0x8062) [0x7f3aa8f89062]
>  6: (clone()+0x6d) [0x7f3aa78c9a3d]
>  NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed
> to interpret this.
>
> --- logging levels ---
>    0/ 5 none
>    0/ 1 lockdep
>    0/ 1 context
>    1/ 1 crush
>    1/ 5 mds
>    1/ 5 mds_balancer
>    1/ 5 mds_locker
>    1/ 5 mds_log
>    1/ 5 mds_log_expire
>    1/ 5 mds_migrator
>    0/ 1 buffer
>    0/ 1 timer
>    0/ 1 filer
>    0/ 1 striper
>    0/ 1 objecter
>    0/ 5 rados
>    0/ 5 rbd
>    0/ 5 journaler
>    0/ 5 objectcacher
>    0/ 5 client
>    0/ 5 osd
>    0/ 5 optracker
>    0/ 5 objclass
>    1/ 3 filestore
>    1/ 3 keyvaluestore
>    1/ 3 journal
>    0/ 5 ms
>    1/ 5 mon
>    0/10 monc
>    1/ 5 paxos
>    0/ 5 tp
>    1/ 5 auth
>    1/ 5 crypto
>    1/ 1 finisher
>    1/ 5 heartbeatmap
>    1/ 5 perfcounter
>    1/ 5 rgw
>    1/ 5 javaclient
>    1/ 5 asok
>    1/ 1 throttle
>   -2/-2 (syslog threshold)
>   -1/-1 (stderr threshold)
>   max_recent     10000
>   max_new         1000
>   log_file /var/log/ceph/ceph-osd.6.log
> --- end dump of recent events ---
> 2014-07-22 01:01:30.352843 7f3aa5c22700 -1 *** Caught signal (Aborted) **
>  in thread 7f3aa5c22700
>
>  ceph version 0.82 (14085f42ddd0fef4e7e1dc99402d07a8df82c04e)
>  1: /usr/bin/ceph-osd() [0xaac562]
>  2: (()+0xf880) [0x7f3aa8f90880]
>  3: (gsignal()+0x39) [0x7f3aa78193a9]
>  4: (abort()+0x148) [0x7f3aa781c4c8]
>  5: (__gnu_cxx::__verbose_terminate_handler()+0x155) [0x7f3aa81065e5]
>  6: (()+0x5e746) [0x7f3aa8104746]
>  7: (()+0x5e773) [0x7f3aa8104773]
>  8: (()+0x5e9b2) [0x7f3aa81049b2]
>  9: (ceph::__ceph_assert_fail(char const*, char const*, int, char
> const*)+0x40a) [0xb85b6a]
>  10: (ceph::HeartbeatMap::_check(ceph::heartbeat_handle_d*, char const*,
> long)+0x2eb) [0xad2cbb]
>  11: (ceph::HeartbeatMap::is_healthy()+0xb6) [0xad34c6]
>  12: (ceph::HeartbeatMap::check_touch_file()+0x28) [0xad3aa8]
>  13: (CephContextServiceThread::entry()+0x13f) [0xb9911f]
>  14: (()+0x8062) [0x7f3aa8f89062]
>  15: (clone()+0x6d) [0x7f3aa78c9a3d]
>  NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed
> to interpret this.
>
> --- begin dump of recent events ---
>      0> 2014-07-22 01:01:30.352843 7f3aa5c22700 -1 *** Caught signal
> (Aborted) **
>  in thread 7f3aa5c22700
>
>  ceph version 0.82 (14085f42ddd0fef4e7e1dc99402d07a8df82c04e)
>  1: /usr/bin/ceph-osd() [0xaac562]
>  2: (()+0xf880) [0x7f3aa8f90880]
>  3: (gsignal()+0x39) [0x7f3aa78193a9]
>  4: (abort()+0x148) [0x7f3aa781c4c8]
>  5: (__gnu_cxx::__verbose_terminate_handler()+0x155) [0x7f3aa81065e5]
>  6: (()+0x5e746) [0x7f3aa8104746]
>  7: (()+0x5e773) [0x7f3aa8104773]
>  8: (()+0x5e9b2) [0x7f3aa81049b2]
>  9: (ceph::__ceph_assert_fail(char const*, char const*, int, char
> const*)+0x40a) [0xb85b6a]
>  10: (ceph::HeartbeatMap::_check(ceph::heartbeat_handle_d*, char const*,
> long)+0x2eb) [0xad2cbb]
>  11: (ceph::HeartbeatMap::is_healthy()+0xb6) [0xad34c6]
>  12: (ceph::HeartbeatMap::check_touch_file()+0x28) [0xad3aa8]
>  13: (CephContextServiceThread::entry()+0x13f) [0xb9911f]
>  14: (()+0x8062) [0x7f3aa8f89062]
>  15: (clone()+0x6d) [0x7f3aa78c9a3d]
>  NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed
> to interpret this.
>
> --- logging levels ---
>    0/ 5 none
>    0/ 1 lockdep
>    0/ 1 context
>    1/ 1 crush
>    1/ 5 mds
>    1/ 5 mds_balancer
>    1/ 5 mds_locker
>    1/ 5 mds_log
>    1/ 5 mds_log_expire
>    1/ 5 mds_migrator
>    0/ 1 buffer
>    0/ 1 timer
>    0/ 1 filer
>    0/ 1 striper
>    0/ 1 objecter
>    0/ 5 rados
>    0/ 5 rbd
>    0/ 5 journaler
>    0/ 5 objectcacher
>    0/ 5 client
>    0/ 5 osd
>    0/ 5 optracker
>    0/ 5 objclass
>    1/ 3 filestore
>    1/ 3 keyvaluestore
>    1/ 3 journal
>    0/ 5 ms
>    1/ 5 mon
>    0/10 monc
>    1/ 5 paxos
>    0/ 5 tp
>    1/ 5 auth
>    1/ 5 crypto
>    1/ 1 finisher
>    1/ 5 heartbeatmap
>    1/ 5 perfcounter
>    1/ 5 rgw
>    1/ 5 javaclient
>    1/ 5 asok
>    1/ 1 throttle
>   -2/-2 (syslog threshold)
>   -1/-1 (stderr threshold)
>   max_recent     10000
>   max_new         1000
>   log_file /var/log/ceph/ceph-osd.6.log
> --- end dump of recent events ---
>
> Regards,
> Hong
>
>
>
>   On Monday, July 21, 2014 9:35 PM, Craig Lewis <clewis at centraldesktop.com>
> wrote:
>
>
>  I'd like to get rid of those inconsistent PGs.  I think fixing those
> will get your MDS working again, but I don't actually know anything about
> MDS.  Still, it's best to work your way up from the bottom.  If the OSDs
> aren't stable, there's no use building services on top of them.
>
>
> It's strange that osd.0 was up, but crashed during deep-scrubbing.  You
> might try disabling deep-scrubs (ceph osd set nodeep-scrub), and see if
> osd.0 will stay up.  If running without deep-scrubbing will get your
> cluster consistent, you can reformat the disk later.
>
> You said osd.2 fails to start, with a corrupt journal error.  There's not
> much you can do there.  You should remove it again, mark it lost, reformat
> the disk, and re-add it to the cluster.
>
>
> I'd rebuild osd.2 first, while leaving osd.0 and osd.1 down.
>
> Do you have enough disk space that osd.2 can take all of the data from
> osd.0 and osd.1?  If so, you can mark osd.0 and osd.1 as DOWN and OUT.  If
> not, make sure that osd.0 and osd.1 are marked DOWN and IN.
>
> Once osd.2 finishes rebuilding, I'd set noin, then bring osd.0 and osd.1
> up.  If they're OUT, that will allow Ceph to copy any unique data they
> might have, but it won't try to write anything to them.  If they're IN,
> well, Ceph will try to write to them.  Either way, I'm hoping that they
> stay up long enough for you to get 100% consistent.
>
>
>
>
>
>
> On Sun, Jul 20, 2014 at 7:01 PM, hjcho616 <hjcho616 at yahoo.com> wrote:
>
>  Based on your suggestion here is what I did.
>
> # ceph osd set nobackfill
> set nobackfill
> # ceph osd set norecovery
> Invalid command:  norecovery not in
> pause|noup|nodown|noout|noin|nobackfill|norecover|noscrub|nodeep-scrub|notieragent
> osd set
> pause|noup|nodown|noout|noin|nobackfill|norecover|noscrub|nodeep-scrub|notieragent
> :  set <key>
> Error EINVAL: invalid command
> # ceph osd set norecover
> set norecover
> # ceph osd set noin
> set noin
> # ceph create osd
> no valid command found; 10 closest matches:
> osd tier remove <poolname> <poolname>
> osd tier cache-mode <poolname> none|writeback|forward|readonly
> osd thrash <int[0-]>
> osd tier add <poolname> <poolname> {--force-nonempty}
> osd pool stats {<name>}
> osd reweight-by-utilization {<int[100-]>}
> osd pool set <poolname>
> size|min_size|crash_replay_interval|pg_num|pgp_num|crush_ruleset|hashpspool|hit_set_type|hit_set_period|hit_set_count|hit_set_fpp|debug_fake_ec_pool|target_max_bytes|target_max_objects|cache_target_dirty_ratio|cache_target_full_ratio|cache_min_flush_age|cache_min_evict_age|auid
> <val> {--yes-i-really-mean-it}
> osd pool set-quota <poolname> max_objects|max_bytes <val>
> osd pool rename <poolname> <poolname>
> osd pool get <poolname>
> size|min_size|crash_replay_interval|pg_num|pgp_num|crush_ruleset|hit_set_type|hit_set_period|hit_set_count|hit_set_fpp|auid
> Error EINVAL: invalid command
> # ceph osd create
> 0
> # ceph osd create
> 1
> # ceph osd create
> 2
> # start ceph-osd id=0
> bash: start: command not found
> # /etc/init.d/ceph start osd.0
> === osd.0 ===
> 2014-07-18 21:21:37.207159 7ff2c64d7700  0 librados: osd.0 authentication
> error (1) Operation not permitted
> Error connecting to cluster: PermissionError
> failed: 'timeout 10 /usr/bin/ceph -c /etc/ceph/ceph.conf --name=osd.0
> --keyring=/var/lib/ceph/osd/ceph-0/keyring osd crush create-or-move -- 0
> 1.82 host=OSD1 root=default'
> # ceph status
>     cluster 9b2c9bca-112e-48b0-86fc-587ef9a52948
>      health HEALTH_ERR 164 pgs degraded; 38 pgs inconsistent; 192 pgs
> stuck unclean; recovery 1484224/3513098 objects degraded (42.248%); 1374
> scrub errors; mds cluster is degraded; mds MDS1 is laggy;
> noin,nobackfill,norecover flag(s) set
>      monmap e1: 1 mons at {MDS1=192.168.1.20:6789/0}, election epoch 1,
> quorum 0 MDS1
>      mdsmap e7182: 1/1/1 up {0=MDS1=up:replay(laggy or crashed)}
>      osdmap e3133: 6 osds: 3 up, 3 in
>             flags noin,nobackfill,norecover
>       pgmap v309437: 192 pgs, 3 pools, 1571 GB data, 1715 kobjects
>             1958 GB used, 3627 GB / 5586 GB avail
>             1484224/3513098 objects degraded (42.248%)
>                  131 active+degraded
>                   23 active+remapped
>                   33 active+degraded+inconsistent
>                    5 active+remapped+inconsistent
> # ceph osd stat
>      osdmap e3133: 6 osds: 3 up, 3 in
>             flags noin,nobackfill,norecover
> # ceph auth get-or-create osd.0 mon 'allow rwx' osd 'allow *' -o
> /var/lib/ceph/osd/ceph-0/keyring
>
> # /etc/init.d/ceph start osd.0
> === osd.0 ===
> create-or-move updating item name 'osd.0' weight 1.82 at location
> {host=OSD1,root=default} to crush map
> Starting Ceph osd.0 on OSD1...
> starting osd.0 at :/0 osd_data /var/lib/ceph/osd/ceph-0
> /var/lib/ceph/osd/ceph-0/journal
> root at OSD1:/home/genie# ceph auth get-or-create osd.1 mon 'allow rwx' osd
> 'allow *' -o /var/lib/ceph/osd/ceph-1/keyring
> root at OSD1:/home/genie# ceph auth get-or-create osd.2 mon 'allow rwx' osd
> 'allow *' -o /var/lib/ceph/osd/ceph-2/keyring
> root at OSD1:/home/genie# /etc/init.d/ceph start osd.1
> === osd.1 ===
> failed: 'timeout 10 /usr/bin/ceph -c /etc/ceph/ceph.conf --name=osd.1
> --keyring=/var/lib/ceph/osd/ceph-1/keyring osd crush create-or-move -- 1
> 1.82 host=OSD1 root=default'
> # /etc/init.d/ceph start osd.2
> === osd.2 ===
> create-or-move updating item name 'osd.2' weight 1.82 at location
> {host=OSD1,root=default} to crush map
> Starting Ceph osd.2 on OSD1...
> starting osd.2 at :/0 osd_data /var/lib/ceph/osd/ceph-2
> /var/lib/ceph/osd/ceph-2/journal
> # /etc/init.d/ceph start osd.1
> === osd.1 ===
> failed: 'timeout 10 /usr/bin/ceph -c /etc/ceph/ceph.conf --name=osd.1
> --keyring=/var/lib/ceph/osd/ceph-1/keyring osd crush create-or-move -- 1
> 1.82 host=OSD1 root=default'
> # ceph health
> Segmentation fault
> # ceph health
> Bus error
> # ceph health
> HEALTH_ERR 164 pgs degraded; 38 pgs inconsistent; 192 pgs stuck unclean;
> recovery 1484224/3513098 objects degraded (42.248%); 1374 scrub errors; mds
> cluster is degraded; mds MDS1 is laggy; noin,nobackfill,norecover flag(s)
> set
> # /etc/init.d/ceph start osd.1
> === osd.1 ===
> create-or-move updating item name 'osd.1' weight 1.82 at location
> {host=OSD1,root=default} to crush map
> Starting Ceph osd.1 on OSD1...
> starting osd.1 at :/0 osd_data /var/lib/ceph/osd/ceph-1
> /var/lib/ceph/osd/ceph-1/journal
> # ceph -w
>     cluster 9b2c9bca-112e-48b0-86fc-587ef9a52948
>      health HEALTH_ERR 164 pgs degraded; 38 pgs inconsistent; 192 pgs
> stuck unclean; recovery 1484224/3513098 objects degraded (42.248%); 1374
> scrub errors; mds cluster is degraded; mds MDS1 is laggy;
> noin,nobackfill,norecover flag(s) set
>      monmap e1: 1 mons at {MDS1=192.168.1.20:6789/0}, election epoch 1,
> quorum 0 MDS1
>      mdsmap e7182: 1/1/1 up {0=MDS1=up:replay(laggy or crashed)}
>      osdmap e3137: 6 osds: 4 up, 3 in
>             flags noin,nobackfill,norecover
>       pgmap v309463: 192 pgs, 3 pools, 1571 GB data, 1715 kobjects
>             1958 GB used, 3627 GB / 5586 GB avail
>             1484224/3513098 objects degraded (42.248%)
>                  131 active+degraded
>                   23 active+remapped
>                   33 active+degraded+inconsistent
>                    5 active+remapped+inconsistent
>
> 2014-07-19 21:34:59.166709 mon.0 [INF] pgmap v309463: 192 pgs: 131
> active+degraded, 23 active+remapped, 33 active+degraded+inconsistent, 5
> active+remapped+inconsistent; 1571 GB data, 1958 GB used, 3627 GB / 5586 GB
> avail; 1484224/3513098 objects degraded (42.248%)
>
>
> osd.2 doesn't come up.  osd.1 uses little memory compared to osd.0, but it
> stays alive.  Killed osd.1 and osd.2 for now.  At this point osd.0's CPU
> was on and off for a while.  But it didn't kill it.  So I did ceph osd
> unset noin and restarted osd.0.  It seemed to be doing something for a long
> time.  I let it run over night.  Found it crashed today.  Below is the log
> of it.
>
>
>    -20> 2014-07-20 00:54:10.924602 7fb562528700  5 -- op tracker -- , seq:
> 4847, time: 2014-07-20 00:54:10.924244, event: header_read, op:
> osd_sub_op(unknown.0.0:0 1.29 0//0//-1 [scrub-reserve] v 0'0
> snapset=0=[]:[] snapc=0=[])
>    -19> 2014-07-20 00:54:10.924652 7fb562528700  5 -- op tracker -- , seq:
> 4847, time: 2014-07-20 00:54:10.924250, event: throttled, op:
> osd_sub_op(unknown.0.0:0 1.29 0//0//-1 [scrub-reserve] v 0'0
> snapset=0=[]:[] snapc=0=[])
>    -18> 2014-07-20 00:54:10.924698 7fb562528700  5 -- op tracker -- , seq:
> 4847, time: 2014-07-20 00:54:10.924458, event: all_read, op:
> osd_sub_op(unknown.0.0:0 1.29 0//0//-1 [scrub-reserve] v 0'0
> snapset=0=[]:[] snapc=0=[])
>    -17> 2014-07-20 00:54:10.924743 7fb562528700  5 -- op tracker -- , seq:
> 4847, time: 0.000000, event: dispatched, op: osd_sub_op(unknown.0.0:0 1.29
> 0//0//-1 [scrub-reserve] v 0'0 snapset=0=[]:[] snapc=0=[])
>    -16> 2014-07-20 00:54:10.924880 7fb54d78f700  5 -- op tracker -- , seq:
> 4847, time: 2014-07-20 00:54:10.924861, event: reached_pg, op:
> osd_sub_op(unknown.0.0:0 1.29 0//0//-1 [scrub-reserve] v 0'0
> snapset=0=[]:[] snapc=0=[])
>    -15> 2014-07-20 00:54:10.924936 7fb54d78f700  5 -- op tracker -- , seq:
> 4847, time: 2014-07-20 00:54:10.924915, event: started, op:
> osd_sub_op(unknown.0.0:0 1.29 0//0//-1 [scrub-reserve] v 0'0
> snapset=0=[]:[] snapc=0=[])
>    -14> 2014-07-20 00:54:10.924974 7fb54d78f700  1 --
> 192.168.2.30:6800/18511 --> 192.168.2.31:6804/13991 --
> osd_sub_op_reply(unknown.0.0:0 1.29 0//0//-1 [scrub-reserve] ack, result =
> 0) v2 -- ?+1 0x10503680 con 0xfdca000
>    -13> 2014-07-20 00:54:10.925053 7fb54d78f700  5 -- op tracker -- , seq:
> 4847, time: 2014-07-20 00:54:10.925034, event: done, op:
> osd_sub_op(unknown.0.0:0 1.29 0//0//-1 [scrub-reserve] v 0'0
> snapset=0=[]:[] snapc=0=[])
>    -12> 2014-07-20 00:54:10.926801 7fb562528700  1 --
> 192.168.2.30:6800/18511 <== osd.4 192.168.2.31:6804/13991 1742 ====
> osd_sub_op(unknown.0.0:0 1.29 0//0//-1 [scrub-unreserve] v 0'0
> snapset=0=[]:[] snapc=0=[]) v10 ==== 1145+0+0 (2357365982 0 0) 0x1045a100
> con 0xfdca000
>    -11> 2014-07-20 00:54:10.926912 7fb562528700  5 -- op tracker -- , seq:
> 4848, time: 2014-07-20 00:54:10.926624, event: header_read, op:
> osd_sub_op(unknown.0.0:0 1.29 0//0//-1 [scrub-unreserve] v 0'0
> snapset=0=[]:[] snapc=0=[])
>    -10> 2014-07-20 00:54:10.926961 7fb562528700  5 -- op tracker -- , seq:
> 4848, time: 2014-07-20 00:54:10.926628, event: throttled, op:
> osd_sub_op(unknown.0.0:0 1.29 0//0//-1 [scrub-unreserve] v 0'0
> snapset=0=[]:[] snapc=0=[])
>     -9> 2014-07-20 00:54:10.927004 7fb562528700  5 -- op tracker -- , seq:
> 4848, time: 2014-07-20 00:54:10.926786, event: all_read, op:
> osd_sub_op(unknown.0.0:0 1.29 0//0//-1 [scrub-unreserve] v 0'0
> snapset=0=[]:[] snapc=0=[])
>     -8> 2014-07-20 00:54:10.927046 7fb562528700  5 -- op tracker -- , seq:
> 4848, time: 0.000000, event: dispatched, op: osd_sub_op(unknown.0.0:0 1.29
> 0//0//-1 [scrub-unreserve] v 0'0 snapset=0=[]:[] snapc=0=[])
>     -7> 2014-07-20 00:54:10.927179 7fb54df90700  5 -- op tracker -- , seq:
> 4848, time: 2014-07-20 00:54:10.927160, event: reached_pg, op:
> osd_sub_op(unknown.0.0:0 1.29 0//0//-1 [scrub-unreserve] v 0'0
> snapset=0=[]:[] snapc=0=[])
>     -6> 2014-07-20 00:54:10.927237 7fb54df90700  5 -- op tracker -- , seq:
> 4848, time: 2014-07-20 00:54:10.927216, event: started, op:
> osd_sub_op(unknown.0.0:0 1.29 0//0//-1 [scrub-unreserve] v 0'0
> snapset=0=[]:[] snapc=0=[])
>     -5> 2014-07-20 00:54:10.927289 7fb54df90700  5 -- op tracker -- , seq:
> 4848, time: 2014-07-20 00:54:10.927269, event: done, op:
> osd_sub_op(unknown.0.0:0 1.29 0//0//-1 [scrub-unreserve] v 0'0
> snapset=0=[]:[] snapc=0=[])
>     -4> 2014-07-20 00:54:10.941372 7fb551f98700  1 --
> 192.168.2.30:6801/18511 <== osd.3 192.168.1.31:0/13623 776 ====
> osd_ping(ping e3144 stamp 2014-07-20 00:54:10.942416) v2 ==== 47+0+0
> (216963345 0 0) 0x103c0e00 con 0x1001bce0
>     -3> 2014-07-20 00:54:10.941451 7fb551f98700  1 --
> 192.168.2.30:6801/18511 --> 192.168.1.31:0/13623 -- osd_ping(ping_reply
> e3144 stamp 2014-07-20 00:54:10.942416) v2 -- ?+0 0x100e2540 con 0x1001bce0
>     -2> 2014-07-20 00:54:10.941742 7fb55379b700  1 --
> 192.168.1.30:6801/18511 <== osd.3 192.168.1.31:0/13623 776 ====
> osd_ping(ping e3144 stamp 2014-07-20 00:54:10.942416) v2 ==== 47+0+0
> (216963345 0 0) 0x10547880 con 0xff8db80
>     -1> 2014-07-20 00:54:10.941842 7fb55379b700  1 --
> 192.168.1.30:6801/18511 --> 192.168.1.31:0/13623 -- osd_ping(ping_reply
> e3144 stamp 2014-07-20 00:54:10.942416) v2 -- ?+0 0x10254a80 con 0xff8db80
>      0> 2014-07-20 00:54:11.646226 7fb54c78d700 -1 os/DBObjectMap.cc: In
> function 'virtual bool DBObjectMap::DBObjectMapIteratorImpl::valid()'
> thread 7fb54c78d700 time 2014-07-20 00:54:11.640719
> os/DBObjectMap.cc: 399: FAILED assert(!valid || cur_iter->valid())
>
>  ceph version 0.82 (14085f42ddd0fef4e7e1dc99402d07a8df82c04e)
>  1: /usr/bin/ceph-osd() [0xa72172]
>  2: (ReplicatedBackend::be_deep_scrub(hobject_t const&, ScrubMap::object&,
> ThreadPool::TPHandle&)+0x6c3) [0xa2df03]
>  3: (PGBackend::be_scan_list(ScrubMap&, std::vector<hobject_t,
> std::allocator<hobject_t> > const&, bool, ThreadPool::TPHandle&)+0x503)
> [0x98c523]
>  4: (PG::build_scrub_map_chunk(ScrubMap&, hobject_t, hobject_t, bool,
> ThreadPool::TPHandle&)+0x10b) [0x891d4b]
>  5: (PG::replica_scrub(MOSDRepScrub*, ThreadPool::TPHandle&)+0x456)
> [0x8925d6]
>  6: (OSD::RepScrubWQ::_process(MOSDRepScrub*,
> ThreadPool::TPHandle&)+0x10a) [0x7b00fa]
>  7: (ThreadPool::worker(ThreadPool::WorkThread*)+0x68a) [0xb7792a]
>  8: (ThreadPool::WorkThread::entry()+0x10) [0xb78b80]
>  9: (()+0x8062) [0x7fb56b184062]
>  10: (clone()+0x6d) [0x7fb569ac4a3d]
>  NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed
> to interpret this.
>
> --- logging levels ---
>    0/ 5 none
>    0/ 1 lockdep
>    0/ 1 context
>    1/ 1 crush
>    1/ 5 mds
>    1/ 5 mds_balancer
>    1/ 5 mds_locker
>    1/ 5 mds_log
>    1/ 5 mds_log_expire
>    1/ 5 mds_migrator
>    0/ 1 buffer
>    0/ 1 timer
>    0/ 1 filer
>    0/ 1 striper
>    0/ 1 objecter
>    0/ 5 rados
>    0/ 5 rbd
>    0/ 5 journaler
>    0/ 5 objectcacher
>    0/ 5 client
>    0/ 5 osd
>    0/ 5 optracker
>    0/ 5 objclass
>    1/ 3 filestore
>    1/ 3 keyvaluestore
>    1/ 3 journal
>    0/ 5 ms
>    1/ 5 mon
>    0/10 monc
>    1/ 5 paxos
>    0/ 5 tp
>    1/ 5 auth
>    1/ 5 crypto
>    1/ 1 finisher
>    1/ 5 heartbeatmap
>    1/ 5 perfcounter
>    1/ 5 rgw
>    1/ 5 javaclient
>    1/ 5 asok
>    1/ 1 throttle
>   -2/-2 (syslog threshold)
>   -1/-1 (stderr threshold)
>   max_recent     10000
>   max_new         1000
>   log_file /var/log/ceph/ceph-osd.0.log
> --- end dump of recent events ---
> 2014-07-20 00:54:11.998700 7fb54c78d700 -1 *** Caught signal (Aborted) **
>  in thread 7fb54c78d700
>
>  ceph version 0.82 (14085f42ddd0fef4e7e1dc99402d07a8df82c04e)
>  1: /usr/bin/ceph-osd() [0xaac562]
>  2: (()+0xf880) [0x7fb56b18b880]
>  3: (gsignal()+0x39) [0x7fb569a143a9]
>  4: (abort()+0x148) [0x7fb569a174c8]
>  5: (__gnu_cxx::__verbose_terminate_handler()+0x155) [0x7fb56a3015e5]
>  6: (()+0x5e746) [0x7fb56a2ff746]
>  7: (()+0x5e773) [0x7fb56a2ff773]
>  8: (()+0x5e9b2) [0x7fb56a2ff9b2]
>  9: (ceph::__ceph_assert_fail(char const*, char const*, int, char
> const*)+0x40a) [0xb85b6a]
>  10: /usr/bin/ceph-osd() [0xa72172]
>
>  11: (ReplicatedBackend::be_deep_scrub(hobject_t const&,
> ScrubMap::object&, ThreadPool::TPHandle&)+0x6c3) [0xa2df03]
>  12: (PGBackend::be_scan_list(ScrubMap&, std::vector<hobject_t,
> std::allocator<hobject_t> > const&, bool, ThreadPool::TPHandle&)+0x503)
> [0x98c523]
>  13: (PG::build_scrub_map_chunk(ScrubMap&, hobject_t, hobject_t, bool,
> ThreadPool::TPHandle&)+0x10b) [0x891d4b]
>  14: (PG::replica_scrub(MOSDRepScrub*, ThreadPool::TPHandle&)+0x456)
> [0x8925d6]
>  15: (OSD::RepScrubWQ::_process(MOSDRepScrub*,
> ThreadPool::TPHandle&)+0x10a) [0x7b00fa]
>  16: (ThreadPool::worker(ThreadPool::WorkThread*)+0x68a) [0xb7792a]
>  17: (ThreadPool::WorkThread::entry()+0x10) [0xb78b80]
>  18: (()+0x8062) [0x7fb56b184062]
>  19: (clone()+0x6d) [0x7fb569ac4a3d]
>  NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed
> to interpret this.
>
> --- begin dump of recent events ---
>     -1> 2014-07-20 00:54:11.755763 7fb565618700  5 osd.0 3144 tick
>      0> 2014-07-20 00:54:11.998700 7fb54c78d700 -1 *** Caught signal
> (Aborted) **
>  in thread 7fb54c78d700
>
>  ceph version 0.82 (14085f42ddd0fef4e7e1dc99402d07a8df82c04e)
>  1: /usr/bin/ceph-osd() [0xaac562]
>  2: (()+0xf880) [0x7fb56b18b880]
>  3: (gsignal()+0x39) [0x7fb569a143a9]
>  4: (abort()+0x148) [0x7fb569a174c8]
>  5: (__gnu_cxx::__verbose_terminate_handler()+0x155) [0x7fb56a3015e5]
>  6: (()+0x5e746) [0x7fb56a2ff746]
>  7: (()+0x5e773) [0x7fb56a2ff773]
>  8: (()+0x5e9b2) [0x7fb56a2ff9b2]
>  9: (ceph::__ceph_assert_fail(char const*, char const*, int, char
> const*)+0x40a) [0xb85b6a]
>  10: /usr/bin/ceph-osd() [0xa72172]
>  11: (ReplicatedBackend::be_deep_scrub(hobject_t const&,
> ScrubMap::object&, ThreadPool::TPHandle&)+0x6c3) [0xa2df03]
>  12: (PGBackend::be_scan_list(ScrubMap&, std::vector<hobject_t,
> std::allocator<hobject_t> > const&, bool, ThreadPool::TPHandle&)+0x503)
> [0x98c523]
>  13: (PG::build_scrub_map_chunk(ScrubMap&, hobject_t, hobject_t, bool,
> ThreadPool::TPHandle&)+0x10b) [0x891d4b]
>  14: (PG::replica_scrub(MOSDRepScrub*, ThreadPool::TPHandle&)+0x456)
> [0x8925d6]
>  15: (OSD::RepScrubWQ::_process(MOSDRepScrub*,
> ThreadPool::TPHandle&)+0x10a) [0x7b00fa]
>  16: (ThreadPool::worker(ThreadPool::WorkThread*)+0x68a) [0xb7792a]
>  17: (ThreadPool::WorkThread::entry()+0x10) [0xb78b80]
>  18: (()+0x8062) [0x7fb56b184062]
>  19: (clone()+0x6d) [0x7fb569ac4a3d]
>  NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed
> to interpret this.
>
> --- logging levels ---
>    0/ 5 none
>    0/ 1 lockdep
>    0/ 1 context
>    1/ 1 crush
>    1/ 5 mds
>    1/ 5 mds_balancer
>    1/ 5 mds_locker
>    1/ 5 mds_log
>    1/ 5 mds_log_expire
>    1/ 5 mds_migrator
>    0/ 1 buffer
>    0/ 1 timer
>    0/ 1 filer
>    0/ 1 striper
>    0/ 1 objecter
>    0/ 5 rados
>    0/ 5 rbd
>    0/ 5 journaler
>    0/ 5 objectcacher
>    0/ 5 client
>    0/ 5 osd
>    0/ 5 optracker
>    0/ 5 objclass
>    1/ 3 filestore
>    1/ 3 keyvaluestore
>    1/ 3 journal
>    0/ 5 ms
>    1/ 5 mon
>    0/10 monc
>    1/ 5 paxos
>    0/ 5 tp
>    1/ 5 auth
>    1/ 5 crypto
>    1/ 1 finisher
>    1/ 5 heartbeatmap
>    1/ 5 perfcounter
>    1/ 5 rgw
>    1/ 5 javaclient
>    1/ 5 asok
>    1/ 1 throttle
>   -2/-2 (syslog threshold)
>   -1/-1 (stderr threshold)
>   max_recent     10000
>   max_new         1000
>   log_file /var/log/ceph/ceph-osd.0.log
> --- end dump of recent events ---
>
>
> What can I do about this one?
>
> Regards,
> Hong
>
>
>
>
>     On Friday, July 18, 2014 5:16 PM, Craig Lewis <
> clewis at centraldesktop.com> wrote:
>
>
>  That I can't help you with.  I'm a pure RadosGW user.  But OSD stability
> affects everybody. :-P
>
>
> On Fri, Jul 18, 2014 at 2:34 PM, hjcho616 <hjcho616 at yahoo.com> wrote:
>
> Thanks Craig.  I will try this soon.  BTW should I upgrade to 0.80.4
> first?  The MDS journal issue seems to be one of the issue I am running
> into.
>
> Regards,
> Hong
>
>
>   On Friday, July 18, 2014 4:14 PM, Craig Lewis <clewis at centraldesktop.com>
> wrote:
>
>
>  If osd.3, osd.4, and osd.5 are stable, your cluster should be working
> again.  What does ceph status say?
>
>
> I was able to re-add removed osd.
> Here's what I did on my dev cluster:
> stop ceph-osd id=0
> ceph osd down 0
> ceph osd out 0
> ceph osd rm 0
> ceph osd crush rm osd.0
>
> Now my osd tree and osd dump do not show osd.0.  The cluster was degraded,
> but did not do any backfilling because I require 3x replication on 3
> different hosts, and Ceph can't satisfy that with 2 osds.
>
> On the same host, I ran:
> ceph osd create        # Returned ID 0
> start ceph-osd id=0
>
>
> osd.0 started up and joined the cluster.  Once peering completed, all of
> the PGs recovered quickly.  I didn't have any writes on the cluster while I
> was doing this.
>
> So it looks like you can just re-create and start those deleted osds.
>
>
>
> In your situation, I would do the following.  Before you start, go through
> this, and make sure you understand all the steps.  Worst case, you can
> always undo this by removing the osds again, and you'll be back to where
> you are now.
>
> ceph osd set nobackfill
> ceph osd set norecovery
> ceph osd set noin
> ceph create osd   # Should return 0.  Abort if it doesn't.
> ceph create osd   # Should return 1.  Abort if it doesn't.
> ceph create osd   # Should return 2.  Abort if it doesn't.
> start ceph-osd id=0
>
> Watch ceph -w and top.  Hopefully ceph-osd id=0 will use some CPU, then
> go UP, and drop to 0% cpu.  If so,
> ceph osd unset noin
> restart ceph-osd id=0
>
> Now osd.0 should go UP and IN, use some CPU for a while, then drop to 0%
> cpu.  If osd.0 drops out now,  set noout, and shut it down.
>
> set noin again, and start osd.1.  When it's stable, do it again for osd.2.
>
> Once as many as possible are up and stable:
> ceph osd unset nobackfill
> ceph osd unset norecovery
>
> Now it should start recovering.  If your osds start dropping out now,  set
> noout, and shut down the ones that are having problems.
>
>
> The goal is to get all the stable osds up, in, and recovered.  Once that's
> done, we can figure out what to do with the unstable osds.
>
>
>
>
>
>
>
>
>
> On Thu, Jul 17, 2014 at 9:29 PM, hjcho616 <hjcho616 at yahoo.com> wrote:
>
> Sorry Craig.  I thought I sent both but second part didn't copy right.
>  For some reason over night MDS and MON decided to stop so I started it
> when I was running those commands.  Interestingly MDS didn't fail at the
> time like it used to.  So I thought something was being fixed?  Then I now
> realize MDS probably couldn't get to the data because OSD were down.  Now
> that I brought up the OSDs MDS crashed again. =P
>
> $ ceph osd tree
> # id    weight  type name       up/down reweight
> -1      5.46    root default
> -2      0               host OSD1
> -3      5.46            host OSD2
> 3       1.82                    osd.3   up      1
> 4       1.82                    osd.4   up      1
> 5       1.82                    osd.5   up      1
>
> $ ceph osd dump
> epoch 3125
> fsid 9b2c9bca-112e-48b0-86fc-587ef9a52948
> created 2014-02-08 01:57:34.086532
> modified 2014-07-17 23:24:10.823596
> flags
> pool 0 'data' replicated size 2 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 64 pgp_num 64 last_change 1 crash_replay_interval 45
> stripe_width 0
> pool 1 'metadata' replicated size 2 min_size 1 crush_ruleset 1 object_hash
> rjenkins pg_num 64 pgp_num 64 last_change 1 stripe_width 0
> pool 2 'rbd' replicated size 2 min_size 1 crush_ruleset 2 object_hash
> rjenkins pg_num 64 pgp_num 64 last_change 1 stripe_width 0
> max_osd 6
> osd.3 up   in  weight 1 up_from 3120 up_thru 3122 down_at 3116
> last_clean_interval [2858,3113) 192.168.1.31:6803/13623
> 192.168.2.31:6802/13623 192.168.2.31:6803/13623 192.168.1.31:6804/13623
> exists,up 4f86a418-6c67-4cb4-83a1-6c123c890036
> osd.4 up   in  weight 1 up_from 3121 up_thru 3122 down_at 3116
> last_clean_interval [2859,3113) 192.168.1.31:6806/13991
> 192.168.2.31:6804/13991 192.168.2.31:6805/13991 192.168.1.31:6807/13991
> exists,up 3d5e3843-7a47-44b0-b276-61c4b1d62900
> osd.5 up   in  weight 1 up_from 3118 up_thru 3118 down_at 3116
> last_clean_interval [2856,3113) 192.168.1.31:6800/13249
> 192.168.2.31:6800/13249 192.168.2.31:6801/13249 192.168.1.31:6801/13249
> exists,up eec86483-2f35-48a4-a154-2eaf26be06b9
> pg_temp 0.2 [4,3]
> pg_temp 0.a [4,5]
> pg_temp 0.c [3,4]
> pg_temp 0.10 [3,4]
> pg_temp 0.15 [3,5]
> pg_temp 0.17 [3,5]
> pg_temp 0.2f [4,5]
> pg_temp 0.3b [4,3]
> pg_temp 0.3c [3,5]
> pg_temp 0.3d [4,5]
> pg_temp 1.1 [4,3]
> pg_temp 1.9 [4,5]
> pg_temp 1.b [3,4]
> pg_temp 1.14 [3,5]
> pg_temp 1.16 [3,5]
> pg_temp 1.2e [4,5]
> pg_temp 1.3a [4,3]
> pg_temp 1.3b [3,5]
> pg_temp 1.3c [4,5]
> pg_temp 2.0 [4,3]
> pg_temp 2.8 [4,5]
> pg_temp 2.a [3,4]
> pg_temp 2.13 [3,5]
> pg_temp 2.15 [3,5]
> pg_temp 2.2d [4,5]
> pg_temp 2.39 [4,3]
> pg_temp 2.3a [3,5]
> pg_temp 2.3b [4,5]
> blacklist 192.168.1.20:6802/30894 expires 2014-07-17 23:48:10.823576
> blacklist 192.168.1.20:6801/30651 expires 2014-07-17 23:47:55.562984
>
> Regards,
> Hong
>
>
>   On Thursday, July 17, 2014 3:30 PM, Craig Lewis <
> clewis at centraldesktop.com> wrote:
>
>
> You gave me 'ceph osd dump' twice.... can I see 'ceph osd tree' too?
>
> Why are osd.3, osd.4, and osd.5 down?
>
>
> On Thu, Jul 17, 2014 at 11:45 AM, hjcho616 <hjcho616 at yahoo.com> wrote:
>
> Thank you for looking at this.  Below are the outputs you requested.
>
> # ceph osd dump
> epoch 3117
> fsid 9b2c9bca-112e-48b0-86fc-587ef9a52948
> created 2014-02-08 01:57:34.086532
> modified 2014-07-16 22:13:04.385914
> flags
> pool 0 'data' replicated size 2 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 64 pgp_num 64 last_change 1 crash_replay_interval 45
> stripe_width 0
> pool 1 'metadata' replicated size 2 min_size 1 crush_ruleset 1 object_hash
> rjenkins pg_num 64 pgp_num 64 last_change 1 stripe_width 0
> pool 2 'rbd' replicated size 2 min_size 1 crush_ruleset 2 object_hash
> rjenkins pg_num 64 pgp_num 64 last_change 1 stripe_width 0
> max_osd 6
> osd.3 down in  weight 1 up_from 2858 up_thru 3040 down_at 3116
> last_clean_interval [2830,2851) 192.168.1.31:6803/5127
> 192.168.2.31:6805/5127 192.168.2.31:6806/5127 192.168.1.31:6805/5127
> exists 4f86a418-6c67-4cb4-83a1-6c123c890036
> osd.4 down in  weight 1 up_from 2859 up_thru 3043 down_at 3116
> last_clean_interval [2835,2849) 192.168.1.31:6807/5310
> 192.168.2.31:6807/5310 192.168.2.31:6808/5310 192.168.1.31:6808/5310
> exists 3d5e3843-7a47-44b0-b276-61c4b1d62900
> osd.5 down in  weight 1 up_from 2856 up_thru 3042 down_at 3116
> last_clean_interval [2837,2853) 192.168.1.31:6800/4969
> 192.168.2.31:6801/4969 192.168.2.31:6804/4969 192.168.1.31:6801/4969
> exists eec86483-2f35-48a4-a154-2eaf26be06b9
>
> # ceph osd dump
> epoch 3117
> fsid 9b2c9bca-112e-48b0-86fc-587ef9a52948
> created 2014-02-08 01:57:34.086532
> modified 2014-07-16 22:13:04.385914
> flags
> pool 0 'data' replicated size 2 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 64 pgp_num 64 last_change 1 crash_replay_interval 45
> stripe_width 0
> pool 1 'metadata' replicated size 2 min_size 1 crush_ruleset 1 object_hash
> rjenkins pg_num 64 pgp_num 64 last_change 1 stripe_width 0
> pool 2 'rbd' replicated size 2 min_size 1 crush_ruleset 2 object_hash
> rjenkins pg_num 64 pgp_num 64 last_change 1 stripe_width 0
> max_osd 6
> osd.3 down in  weight 1 up_from 2858 up_thru 3040 down_at 3116
> last_clean_interval [2830,2851) 192.168.1.31:6803/5127
> 192.168.2.31:6805/5127 192.168.2.31:6806/5127 192.168.1.31:6805/5127
> exists 4f86a418-6c67-4cb4-83a1-6c123c890036
> osd.4 down in  weight 1 up_from 2859 up_thru 3043 down_at 3116
> last_clean_interval [2835,2849) 192.168.1.31:6807/5310
> 192.168.2.31:6807/5310 192.168.2.31:6808/5310 192.168.1.31:6808/5310
> exists 3d5e3843-7a47-44b0-b276-61c4b1d62900
> osd.5 down in  weight 1 up_from 2856 up_thru 3042 down_at 3116
> last_clean_interval [2837,2853) 192.168.1.31:6800/4969
> 192.168.2.31:6801/4969 192.168.2.31:6804/4969 192.168.1.31:6801/4969
> exists eec86483-2f35-48a4-a154-2eaf26be06b9
>
> Regards,
> Hong
>
>
>
>   On Thursday, July 17, 2014 12:02 PM, Craig Lewis <
> clewis at centraldesktop.com> wrote:
>
>
> I don't believe you can re-add an OSD after `ceph osd rm`, but it's worth
> a shot.  Let me see what I can do on my dev cluster.
>
> What does `ceph osd dump` and `ceph osd tree` say?  I want to make sure
> I'm starting from the same point you are.
>
>
>
> On Wed, Jul 16, 2014 at 7:39 PM, hjcho616 <hjcho616 at yahoo.com> wrote:
>
> I did a "ceph osd rm" for all three but I didn't do anything else to it
> afterwards.  Can this be added back?
>
> Regards,
> Hong
>
>
>   On Wednesday, July 16, 2014 6:54 PM, Craig Lewis <
> clewis at centraldesktop.com> wrote:
>
>
>  For some reason you ended up in my spam folder.  That might be why you
> didn't get any responses.
>
>
> Have you destroyed osd.0, osd.1, and osd.2?  If not, try bringing them up
> one a time.  You might have just one bad disk, which is much better than
> 50% of your disks.
>
> How is the ceph-osd process behaving when it hits the suicide timeout?  I
> had some problems a while back where the ceph-osd process would startup,
> start consuming ~200% CPU for a while, then get stuck using almost exactly
> 100% CPU.  It would get kicked out of the cluster for being unresponsive,
> then suicide.  Repeat.  If that's happening here, I can suggest some things
> to try.
>
>
>
>
>
> On Fri, Jul 11, 2014 at 9:12 PM, hjcho616 <hjcho616 at yahoo.com> wrote:
>
> I have 2 OSD machines with 3 OSD running on each.  One MDS server with 3
> daemons running.  Ran cephfs mostly on 0.78.  One night we lost power for
> split second.  MDS1 and OSD2 went down, OSD1 seemed OK, well turns out OSD1
> suffered most.  Those two machines rebooted and seemed ok except it had
> some inconsistencies.  I waited for a while, didn't fix itself.  So I
> issued 'ceph pg repair pgnum'.  It would try some and some OSD would crash.
>  Tried this for multiple days.  Got some PGs fixed... but mostly it would
> crash an OSD and stop recovering.  dmesg shows something like below.
>
>
>
> [  740.059498] traps: ceph-osd[5279] general protection ip:7f84e75ec75e
> sp:7fff00045bc0 error:0 in libtcmalloc.so.4.1.0[7f84e75b3000+4a000]
>
> and ceph osd log shows something like this.
>
>      -2> 2014-07-09 20:51:01.163571 7fe0f4617700  1 heartbeat_map
> is_healthy 'FileStore::op_tp thread 0x7fe0e8e91700' had timed out after 60
>     -1> 2014-07-09 20:51:01.163609 7fe0f4617700  1 heartbeat_map
> is_healthy 'FileStore::op_tp thread 0x7fe0e8e91700' had suicide timed out
> after 180
>      0> 2014-07-09 20:51:01.169542 7fe0f4617700 -1 common/HeartbeatMap.cc:
> In function 'bool ceph::HeartbeatMap::_check(ceph::heartbeat_handle_d*,
> const char*, time_t)' thread 7fe0f4617700 time 2014-07-09 20:51:01.163642
> common/HeartbeatMap.cc: 79: FAILED assert(0 == "hit suicide timeout")
>
>  ceph version 0.82 (14085f42ddd0fef4e7e1dc99402d07a8df82c04e)
>  1: (ceph::HeartbeatMap::_check(ceph::heartbeat_handle_d*, char const*,
> long)+0x2eb) [0xad2cbb]
>  2: (ceph::HeartbeatMap::is_healthy()+0xb6) [0xad34c6]
>  3: (ceph::HeartbeatMap::check_touch_file()+0x28) [0xad3aa8]
>  4: (CephContextServiceThread::entry()+0x13f) [0xb9911f]
>  5: (()+0x8062) [0x7fe0f797e062]
>  6: (clone()+0x6d) [0x7fe0f62bea3d]
>  NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed
> to interpret this.
>
> --- logging levels ---
>    0/ 5 none
>    0/ 1 lockdep
>    0/ 1 context
>    1/ 1 crush
>    1/ 5 mds
>    1/ 5 mds_balancer
>    1/ 5 mds_locker
>    1/ 5 mds_log
>    1/ 5 mds_log_expire
>    1/ 5 mds_migrator
>    0/ 1 buffer
>    0/ 1 timer
>    0/ 1 filer
>    0/ 1 striper
>    0/ 1 objecter
>    0/ 5 rados
>    0/ 5 rbd
>    0/ 5 journaler
>    0/ 5 objectcacher
>    0/ 5 client
>    0/ 5 osd
>    0/ 5 optracker
>    0/ 5 objclass
>    1/ 3 filestore
>    1/ 3 keyvaluestore
>    1/ 3 journal
>    0/ 5 ms
>    1/ 5 mon
>    0/10 monc
>    1/ 5 paxos
>    0/ 5 tp
>    1/ 5 auth
>    1/ 5 crypto
>    1/ 1 finisher
>    1/ 5 heartbeatmap
>    1/ 5 perfcounter
>    1/ 5 rgw
>    1/ 5 javaclient
>    1/ 5 asok
>    1/ 1 throttle
>   -2/-2 (syslog threshold)
>   -1/-1 (stderr threshold)
>   max_recent     10000
>   max_new         1000
>   log_file /var/log/ceph/ceph-osd.0.log
> --- end dump of recent events ---
> 2014-07-09 20:51:01.534706 7fe0f4617700 -1 *** Caught signal (Aborted) **
>  in thread 7fe0f4617700
>
>
>  ceph version 0.82 (14085f42ddd0fef4e7e1dc99402d07a8df82c04e)
>  1: /usr/bin/ceph-osd() [0xaac562]
>  2: (()+0xf880) [0x7fe0f7985880]
>  3: (gsignal()+0x39) [0x7fe0f620e3a9]
>  4: (abort()+0x148) [0x7fe0f62114c8]
>  5: (__gnu_cxx::__verbose_terminate_handler()+0x155) [0x7fe0f6afb5e5]
>  6: (()+0x5e746) [0x7fe0f6af9746]
>  7: (()+0x5e773) [0x7fe0f6af9773]
>  8: (()+0x5e9b2) [0x7fe0f6af99b2]
>  9: (ceph::__ceph_assert_fail(char const*, char const*, int, char
> const*)+0x40a) [0xb85b6a]
>  10: (ceph::HeartbeatMap::_check(ceph::heartbeat_handle_d*, char const*,
> long)+0x2eb) [0xad2cbb]
>  11: (ceph::HeartbeatMap::is_healthy()+0xb6) [0xad34c6]
>  12: (ceph::HeartbeatMap::check_touch_file()+0x28) [0xad3aa8]
>  13: (CephContextServiceThread::entry()+0x13f) [0xb9911f]
>  14: (()+0x8062) [0x7fe0f797e062]
>  15: (clone()+0x6d) [0x7fe0f62bea3d]
>  NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed
> to interpret this.
>
> --- begin dump of recent events ---
>      0> 2014-07-09 20:51:01.534706 7fe0f4617700 -1 *** Caught signal
> (Aborted) **
>  in thread 7fe0f4617700
>
>  ceph version 0.82 (14085f42ddd0fef4e7e1dc99402d07a8df82c04e)
>  1: /usr/bin/ceph-osd() [0xaac562]
>  2: (()+0xf880) [0x7fe0f7985880]
>  3: (gsignal()+0x39) [0x7fe0f620e3a9]
>  4: (abort()+0x148) [0x7fe0f62114c8]
>  5: (__gnu_cxx::__verbose_terminate_handler()+0x155) [0x7fe0f6afb5e5]
>  6: (()+0x5e746) [0x7fe0f6af9746]
>  7: (()+0x5e773) [0x7fe0f6af9773]
>  8: (()+0x5e9b2) [0x7fe0f6af99b2]
>  9: (ceph::__ceph_assert_fail(char const*, char const*, int, char
> const*)+0x40a) [0xb85b6a]
>  10: (ceph::HeartbeatMap::_check(ceph::heartbeat_handle_d*, char const*,
> long)+0x2eb) [0xad2cbb]
>  11: (ceph::HeartbeatMap::is_healthy()+0xb6) [0xad34c6]
>  12: (ceph::HeartbeatMap::check_touch_file()+0x28) [0xad3aa8]
>
>  ...
>
> [Message clipped]
>
>
>
>
>
>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140812/427d7334/attachment.htm>


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux