Yes, osd.10 is near full because of bad data repartition (not enought PG I suppose), and the difficulty to remove snapshot without overloading the cluster. The problem on osd.25 was a crash during scrub... I tried to reweight it, and set it out, without any success. I have added some OSD too. Logs from my emails «scrub shutdown the OSD process» (the 15th april) : 0> 2013-04-15 09:29:53.708141 7f5a8e3cc700 -1 *** Caught signal (Aborted) ** in thread 7f5a8e3cc700 ceph version 0.56.4-4-gd89ab0e (d89ab0ea6fa8d0961cad82f6a81eccbd3bbd3f55) 1: /usr/bin/ceph-osd() [0x7a6289] 2: (()+0xeff0) [0x7f5aa08faff0] 3: (gsignal()+0x35) [0x7f5a9f3841b5] 4: (abort()+0x180) [0x7f5a9f386fc0] 5: (__gnu_cxx::__verbose_terminate_handler()+0x115) [0x7f5a9fc18dc5] 6: (()+0xcb166) [0x7f5a9fc17166] 7: (()+0xcb193) [0x7f5a9fc17193] 8: (()+0xcb28e) [0x7f5a9fc1728e] 9: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x7c9) [0x8f9549] 10: (ReplicatedPG::_scrub(ScrubMap&)+0x1a78) [0x57a038] 11: (PG::scrub_compare_maps()+0xeb8) [0x696c18] 12: (PG::chunky_scrub()+0x2d9) [0x6c37f9] 13: (PG::scrub()+0x145) [0x6c4e55] 14: (OSD::ScrubWQ::_process(PG*)+0xc) [0x64048c] 15: (ThreadPool::worker(ThreadPool::WorkThread*)+0x879) [0x815179] 16: (ThreadPool::WorkThread::entry()+0x10) [0x817980] 17: (()+0x68ca) [0x7f5aa08f28ca] 18: (clone()+0x6d) [0x7f5a9f421b6d] NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this. --- logging levels --- 0/ 5 none 0/ 1 lockdep 0/ 1 context 1/ 1 crush 1/ 5 mds 1/ 5 mds_balancer 1/ 5 mds_locker 1/ 5 mds_log 1/ 5 mds_log_expire 1/ 5 mds_migrator 0/ 1 buffer 0/ 1 timer 0/ 1 filer 0/ 1 striper 0/ 1 objecter 0/ 5 rados 0/ 5 rbd 0/ 5 journaler 0/ 5 objectcacher 0/ 5 client 0/ 5 osd 0/ 5 optracker 0/ 5 objclass 1/ 3 filestore 1/ 3 journal 0/ 5 ms 1/ 5 mon 0/10 monc 0/ 5 paxos 0/ 5 tp 1/ 5 auth 1/ 5 crypto 1/ 1 finisher 1/ 5 heartbeatmap 1/ 5 perfcounter 1/ 5 rgw 1/ 5 hadoop 1/ 5 javaclient 1/ 5 asok 1/ 1 throttle -1/-1 (syslog threshold) -1/-1 (stderr threshold) max_recent 10000 max_new 1000 log_file /var/log/ceph/osd.25.log --- end dump of recent events --- But now, when I start the osd.25, I obtain : 2013-05-14 08:13:37.304452 7fa0ea4e7700 0 -- :/31236 >> 10.0.0.6:6789/0 pipe(0x22fe470 sd=4 :0 s=1 pgs=0 cs=0 l=1).fault 2013-05-14 08:13:40.361570 7ff469167780 0 ceph version 0.61-11-g3b94f03 (3b94f03ec58abe3d7a6d0359ff9b4d75826f3777), process ceph-osd, pid 31424 2013-05-14 08:13:40.403846 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount FIEMAP ioctl is supported and appears to work 2013-05-14 08:13:40.403873 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount FIEMAP ioctl is disabled via 'filestore fiemap' config option 2013-05-14 08:13:40.404296 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount did NOT detect btrfs 2013-05-14 08:13:40.405753 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount syscall(__NR_syncfs, fd) fully supported 2013-05-14 08:13:40.410164 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount found snaps <> 2013-05-14 08:13:40.511206 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount: enabling WRITEAHEAD journal mode: btrfs not detected 2013-05-14 08:13:43.146895 7ff469167780 1 journal _open /dev/sdi3 fd 18: 26843545600 bytes, block size 4096 bytes, directio = 1, aio = 1 2013-05-14 08:13:43.171846 7ff469167780 1 journal _open /dev/sdi3 fd 18: 26843545600 bytes, block size 4096 bytes, directio = 1, aio = 1 2013-05-14 08:13:43.172772 7ff469167780 1 journal close /dev/sdi3 2013-05-14 08:13:43.179918 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount FIEMAP ioctl is supported and appears to work 2013-05-14 08:13:43.179929 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount FIEMAP ioctl is disabled via 'filestore fiemap' config option 2013-05-14 08:13:43.180340 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount did NOT detect btrfs 2013-05-14 08:13:43.182242 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount syscall(__NR_syncfs, fd) fully supported 2013-05-14 08:13:43.182304 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount found snaps <> 2013-05-14 08:13:43.185621 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount: enabling WRITEAHEAD journal mode: btrfs not detected 2013-05-14 08:13:43.193318 7ff469167780 1 journal _open /dev/sdi3 fd 26: 26843545600 bytes, block size 4096 bytes, directio = 1, aio = 1 2013-05-14 08:13:43.216180 7ff469167780 1 journal _open /dev/sdi3 fd 26: 26843545600 bytes, block size 4096 bytes, directio = 1, aio = 1 2013-05-14 08:13:43.253760 7ff469167780 0 osd.25 58876 crush map has features 262144, adjusting msgr requires for clients 2013-05-14 08:13:43.253776 7ff469167780 0 osd.25 58876 crush map has features 262144, adjusting msgr requires for osds 2013-05-14 08:13:51.675152 7ff450075700 -1 osd/OSD.cc: In function 'OSDMapRef OSDService::get_map(epoch_t)' thread 7ff450075700 time 2013-05-14 08:13:51.661081 osd/OSD.cc: 4838: FAILED assert(_get_map_bl(epoch, bl)) ceph version 0.61-11-g3b94f03 (3b94f03ec58abe3d7a6d0359ff9b4d75826f3777) 1: (OSDService::get_map(unsigned int)+0x918) [0x608268] 2: (OSD::advance_pg(unsigned int, PG*, ThreadPool::TPHandle&, PG::RecoveryCtx*, std::set<boost::intrusive_ptr<PG>, std::less<boost::intrusive_ptr<PG> >, std::allocator<boost::intrusive_ptr<PG> > >*)+0x195) [0x60b725] 3: (OSD::process_peering_events(std::list<PG*, std::allocator<PG*> > const&, ThreadPool::TPHandle&)+0x24b) [0x630c1b] 4: (OSD::PeeringWQ::_process(std::list<PG*, std::allocator<PG*> > const&, ThreadPool::TPHandle&)+0x16) [0x6650b6] 5: (ThreadPool::worker(ThreadPool::WorkThread*)+0x561) [0x813a11] 6: (ThreadPool::WorkThread::entry()+0x10) [0x816530] 7: (()+0x68ca) [0x7ff4686028ca] 8: (clone()+0x6d) [0x7ff466ee0b6d] NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this. --- begin dump of recent events --- -495> 2013-05-14 08:13:40.355375 7ff469167780 5 asok(0x1fbe000) register_command perfcounters_dump hook 0x1fb2010 -494> 2013-05-14 08:13:40.355404 7ff469167780 5 asok(0x1fbe000) register_command 1 hook 0x1fb2010 -493> 2013-05-14 08:13:40.355407 7ff469167780 5 asok(0x1fbe000) register_command perf dump hook 0x1fb2010 -492> 2013-05-14 08:13:40.355419 7ff469167780 5 asok(0x1fbe000) register_command perfcounters_schema hook 0x1fb2010 -491> 2013-05-14 08:13:40.355427 7ff469167780 5 asok(0x1fbe000) register_command 2 hook 0x1fb2010 -490> 2013-05-14 08:13:40.355439 7ff469167780 5 asok(0x1fbe000) register_command perf schema hook 0x1fb2010 -489> 2013-05-14 08:13:40.355442 7ff469167780 5 asok(0x1fbe000) register_command config show hook 0x1fb2010 -488> 2013-05-14 08:13:40.355445 7ff469167780 5 asok(0x1fbe000) register_command config set hook 0x1fb2010 -487> 2013-05-14 08:13:40.355455 7ff469167780 5 asok(0x1fbe000) register_command log flush hook 0x1fb2010 -486> 2013-05-14 08:13:40.355458 7ff469167780 5 asok(0x1fbe000) register_command log dump hook 0x1fb2010 -485> 2013-05-14 08:13:40.355461 7ff469167780 5 asok(0x1fbe000) register_command log reopen hook 0x1fb2010 -484> 2013-05-14 08:13:40.361570 7ff469167780 0 ceph version 0.61-11-g3b94f03 (3b94f03ec58abe3d7a6d0359ff9b4d75826f3777), process ceph-osd, pid 31424 -483> 2013-05-14 08:13:40.393749 7ff469167780 1 -- 10.0.0.3:0/0 learned my addr 10.0.0.3:0/0 -482> 2013-05-14 08:13:40.393764 7ff469167780 1 accepter.accepter.bind my_inst.addr is 10.0.0.3:6806/31424 need_addr=0 -481> 2013-05-14 08:13:40.393789 7ff469167780 1 -- 192.168.42.3:0/0 learned my addr 192.168.42.3:0/0 -480> 2013-05-14 08:13:40.393795 7ff469167780 1 accepter.accepter.bind my_inst.addr is 192.168.42.3:6810/31424 need_addr=0 -479> 2013-05-14 08:13:40.393815 7ff469167780 1 -- 192.168.42.3:0/0 learned my addr 192.168.42.3:0/0 -478> 2013-05-14 08:13:40.393820 7ff469167780 1 accepter.accepter.bind my_inst.addr is 192.168.42.3:6811/31424 need_addr=0 -477> 2013-05-14 08:13:40.394532 7ff469167780 1 finished global_init_daemonize -476> 2013-05-14 08:13:40.397912 7ff469167780 5 asok(0x1fbe000) init /var/run/ceph/ceph-osd.25.asok -475> 2013-05-14 08:13:40.397942 7ff469167780 5 asok(0x1fbe000) bind_and_listen /var/run/ceph/ceph-osd.25.asok -474> 2013-05-14 08:13:40.397977 7ff469167780 5 asok(0x1fbe000) register_command 0 hook 0x1fb10b8 -473> 2013-05-14 08:13:40.397984 7ff469167780 5 asok(0x1fbe000) register_command version hook 0x1fb10b8 -472> 2013-05-14 08:13:40.397991 7ff469167780 5 asok(0x1fbe000) register_command git_version hook 0x1fb10b8 -471> 2013-05-14 08:13:40.397994 7ff469167780 5 asok(0x1fbe000) register_command help hook 0x1fb20d0 -470> 2013-05-14 08:13:40.398100 7ff464ca2700 5 asok(0x1fbe000) entry start -469> 2013-05-14 08:13:40.403846 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount FIEMAP ioctl is supported and appears to work -468> 2013-05-14 08:13:40.403873 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount FIEMAP ioctl is disabled via 'filestore fiemap' config option -467> 2013-05-14 08:13:40.404296 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount did NOT detect btrfs -466> 2013-05-14 08:13:40.405753 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount syscall(__NR_syncfs, fd) fully supported -465> 2013-05-14 08:13:40.410164 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount found snaps <> -464> 2013-05-14 08:13:40.511206 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount: enabling WRITEAHEAD journal mode: btrfs not detected -463> 2013-05-14 08:13:43.144294 7ff469167780 2 journal open /dev/sdi3 fsid e82dd661-2ad1-4103-92d4-26a3926d524b fs_op_seq 1524687 -462> 2013-05-14 08:13:43.146895 7ff469167780 1 journal _open /dev/sdi3 fd 18: 26843545600 bytes, block size 4096 bytes, directio = 1, aio = 1 -461> 2013-05-14 08:13:43.148887 7ff469167780 2 journal read_entry 3012567040 : seq 1524509 111324 bytes -460> 2013-05-14 08:13:43.148986 7ff469167780 2 journal read_entry 3012571136 : seq 1524510 127 bytes -459> 2013-05-14 08:13:43.148999 7ff469167780 2 journal read_entry 3012575232 : seq 1524511 55 bytes -458> 2013-05-14 08:13:43.149007 7ff469167780 2 journal read_entry 3012579328 : seq 1524512 127 bytes -457> 2013-05-14 08:13:43.149014 7ff469167780 2 journal read_entry 3012583424 : seq 1524513 55 bytes -456> 2013-05-14 08:13:43.149021 7ff469167780 2 journal read_entry 3012587520 : seq 1524514 128 bytes -455> 2013-05-14 08:13:43.149955 7ff469167780 2 journal read_entry 3012591616 : seq 1524515 56 bytes -454> 2013-05-14 08:13:43.149976 7ff469167780 2 journal read_entry 3012595712 : seq 1524516 128 bytes -453> 2013-05-14 08:13:43.149984 7ff469167780 2 journal read_entry 3012599808 : seq 1524517 56 bytes -452> 2013-05-14 08:13:43.149991 7ff469167780 2 journal read_entry 3012603904 : seq 1524518 128 bytes -451> 2013-05-14 08:13:43.149998 7ff469167780 2 journal read_entry 3012608000 : seq 1524519 56 bytes -450> 2013-05-14 08:13:43.150006 7ff469167780 2 journal read_entry 3012612096 : seq 1524520 128 bytes -449> 2013-05-14 08:13:43.150015 7ff469167780 2 journal read_entry 3012616192 : seq 1524521 56 bytes -448> 2013-05-14 08:13:43.150023 7ff469167780 2 journal read_entry 3012620288 : seq 1524522 128 bytes -447> 2013-05-14 08:13:43.150031 7ff469167780 2 journal read_entry 3012624384 : seq 1524523 56 bytes -446> 2013-05-14 08:13:43.150050 7ff469167780 2 journal read_entry 3012628480 : seq 1524524 128 bytes -445> 2013-05-14 08:13:43.150062 7ff469167780 2 journal read_entry 3012632576 : seq 1524525 56 bytes -444> 2013-05-14 08:13:43.150086 7ff469167780 2 journal read_entry 3012636672 : seq 1524526 3327 bytes -443> 2013-05-14 08:13:43.150104 7ff469167780 2 journal read_entry 3012640768 : seq 1524527 2067 bytes -442> 2013-05-14 08:13:43.151434 7ff469167780 2 journal read_entry 3012833280 : seq 1524528 192260 bytes -441> 2013-05-14 08:13:43.151454 7ff469167780 2 journal read_entry 3012837376 : seq 1524529 39 bytes -440> 2013-05-14 08:13:43.151470 7ff469167780 2 journal read_entry 3012841472 : seq 1524530 798 bytes -439> 2013-05-14 08:13:43.152603 7ff469167780 2 journal read_entry 3013038080 : seq 1524531 194372 bytes -438> 2013-05-14 08:13:43.152624 7ff469167780 2 journal read_entry 3013042176 : seq 1524532 39 bytes -437> 2013-05-14 08:13:43.152633 7ff469167780 2 journal read_entry 3013046272 : seq 1524533 798 bytes -436> 2013-05-14 08:13:43.153118 7ff469167780 2 journal read_entry 3013201920 : seq 1524534 152708 bytes -435> 2013-05-14 08:13:43.153137 7ff469167780 2 journal read_entry 3013206016 : seq 1524535 39 bytes -434> 2013-05-14 08:13:43.153153 7ff469167780 2 journal read_entry 3013210112 : seq 1524536 2380 bytes -433> 2013-05-14 08:13:43.153686 7ff469167780 2 journal read_entry 3013369856 : seq 1524537 158184 bytes -432> 2013-05-14 08:13:43.153708 7ff469167780 2 journal read_entry 3013373952 : seq 1524538 127 bytes -431> 2013-05-14 08:13:43.154098 7ff469167780 2 journal read_entry 3013378048 : seq 1524539 55 bytes -430> 2013-05-14 08:13:43.154126 7ff469167780 2 journal read_entry 3013382144 : seq 1524540 127 bytes -429> 2013-05-14 08:13:43.154134 7ff469167780 2 journal read_entry 3013386240 : seq 1524541 55 bytes -428> 2013-05-14 08:13:43.154141 7ff469167780 2 journal read_entry 3013390336 : seq 1524542 128 bytes -427> 2013-05-14 08:13:43.154148 7ff469167780 2 journal read_entry 3013394432 : seq 1524543 56 bytes -426> 2013-05-14 08:13:43.154155 7ff469167780 2 journal read_entry 3013398528 : seq 1524544 128 bytes -425> 2013-05-14 08:13:43.154162 7ff469167780 2 journal read_entry 3013402624 : seq 1524545 56 bytes -424> 2013-05-14 08:13:43.154169 7ff469167780 2 journal read_entry 3013406720 : seq 1524546 128 bytes -423> 2013-05-14 08:13:43.154176 7ff469167780 2 journal read_entry 3013410816 : seq 1524547 56 bytes -422> 2013-05-14 08:13:43.154190 7ff469167780 2 journal read_entry 3013414912 : seq 1524548 128 bytes -421> 2013-05-14 08:13:43.154200 7ff469167780 2 journal read_entry 3013419008 : seq 1524549 56 bytes -420> 2013-05-14 08:13:43.154211 7ff469167780 2 journal read_entry 3013423104 : seq 1524550 128 bytes -419> 2013-05-14 08:13:43.154221 7ff469167780 2 journal read_entry 3013427200 : seq 1524551 56 bytes -418> 2013-05-14 08:13:43.154228 7ff469167780 2 journal read_entry 3013431296 : seq 1524552 128 bytes -417> 2013-05-14 08:13:43.154235 7ff469167780 2 journal read_entry 3013435392 : seq 1524553 56 bytes -416> 2013-05-14 08:13:43.154245 7ff469167780 2 journal read_entry 3013439488 : seq 1524554 128 bytes -415> 2013-05-14 08:13:43.154254 7ff469167780 2 journal read_entry 3013443584 : seq 1524555 56 bytes -414> 2013-05-14 08:13:43.154274 7ff469167780 2 journal read_entry 3013447680 : seq 1524556 3812 bytes -413> 2013-05-14 08:13:43.154290 7ff469167780 2 journal read_entry 3013451776 : seq 1524557 1969 bytes -412> 2013-05-14 08:13:43.154729 7ff469167780 2 journal read_entry 3013562368 : seq 1524558 107588 bytes -411> 2013-05-14 08:13:43.154755 7ff469167780 2 journal read_entry 3013566464 : seq 1524559 39 bytes -410> 2013-05-14 08:13:43.154767 7ff469167780 2 journal read_entry 3013570560 : seq 1524560 3641 bytes -409> 2013-05-14 08:13:43.155250 7ff469167780 2 journal read_entry 3013697536 : seq 1524561 126404 bytes -408> 2013-05-14 08:13:43.155270 7ff469167780 2 journal read_entry 3013701632 : seq 1524562 39 bytes -407> 2013-05-14 08:13:43.155280 7ff469167780 2 journal read_entry 3013705728 : seq 1524563 2004 bytes -406> 2013-05-14 08:13:43.155864 7ff469167780 2 journal read_entry 3013840896 : seq 1524564 134672 bytes -405> 2013-05-14 08:13:43.155884 7ff469167780 2 journal read_entry 3013844992 : seq 1524565 128 bytes -404> 2013-05-14 08:13:43.155892 7ff469167780 2 journal read_entry 3013849088 : seq 1524566 56 bytes -403> 2013-05-14 08:13:43.155900 7ff469167780 2 journal read_entry 3013853184 : seq 1524567 524 bytes -402> 2013-05-14 08:13:43.155909 7ff469167780 2 journal read_entry 3013857280 : seq 1524568 2145 bytes -401> 2013-05-14 08:13:43.157004 7ff469167780 2 journal read_entry 3014049792 : seq 1524569 192272 bytes -400> 2013-05-14 08:13:43.157026 7ff469167780 2 journal read_entry 3014053888 : seq 1524570 127 bytes -399> 2013-05-14 08:13:43.157033 7ff469167780 2 journal read_entry 3014057984 : seq 1524571 55 bytes -398> 2013-05-14 08:13:43.157041 7ff469167780 2 journal read_entry 3014062080 : seq 1524572 127 bytes -397> 2013-05-14 08:13:43.157057 7ff469167780 2 journal read_entry 3014066176 : seq 1524573 55 bytes -396> 2013-05-14 08:13:43.157067 7ff469167780 2 journal read_entry 3014070272 : seq 1524574 128 bytes -395> 2013-05-14 08:13:43.157076 7ff469167780 2 journal read_entry 3014074368 : seq 1524575 56 bytes -394> 2013-05-14 08:13:43.157086 7ff469167780 2 journal read_entry 3014078464 : seq 1524576 128 bytes -393> 2013-05-14 08:13:43.157105 7ff469167780 2 journal read_entry 3014082560 : seq 1524577 56 bytes -392> 2013-05-14 08:13:43.157121 7ff469167780 2 journal read_entry 3014086656 : seq 1524578 128 bytes -391> 2013-05-14 08:13:43.157131 7ff469167780 2 journal read_entry 3014090752 : seq 1524579 56 bytes -390> 2013-05-14 08:13:43.157141 7ff469167780 2 journal read_entry 3014094848 : seq 1524580 128 bytes -389> 2013-05-14 08:13:43.157150 7ff469167780 2 journal read_entry 3014098944 : seq 1524581 56 bytes -388> 2013-05-14 08:13:43.157161 7ff469167780 2 journal read_entry 3014103040 : seq 1524582 128 bytes -387> 2013-05-14 08:13:43.157171 7ff469167780 2 journal read_entry 3014107136 : seq 1524583 56 bytes -386> 2013-05-14 08:13:43.157188 7ff469167780 2 journal read_entry 3014111232 : seq 1524584 2930 bytes -385> 2013-05-14 08:13:43.157200 7ff469167780 2 journal read_entry 3014115328 : seq 1524585 798 bytes -384> 2013-05-14 08:13:43.158161 7ff469167780 2 journal read_entry 3014307840 : seq 1524586 192260 bytes -383> 2013-05-14 08:13:43.158181 7ff469167780 2 journal read_entry 3014311936 : seq 1524587 39 bytes -382> 2013-05-14 08:13:43.158192 7ff469167780 2 journal read_entry 3014316032 : seq 1524588 798 bytes -381> 2013-05-14 08:13:43.158269 7ff469167780 2 journal read_entry 3014373376 : seq 1524589 57092 bytes -380> 2013-05-14 08:13:43.158279 7ff469167780 2 journal read_entry 3014377472 : seq 1524590 39 bytes -379> 2013-05-14 08:13:43.158291 7ff469167780 2 journal read_entry 3014381568 : seq 1524591 2106 bytes -378> 2013-05-14 08:13:43.159486 7ff469167780 2 journal read_entry 3014578176 : seq 1524592 192452 bytes -377> 2013-05-14 08:13:43.159506 7ff469167780 2 journal read_entry 3014582272 : seq 1524593 127 bytes -376> 2013-05-14 08:13:43.159516 7ff469167780 2 journal read_entry 3014586368 : seq 1524594 55 bytes -375> 2013-05-14 08:13:43.159527 7ff469167780 2 journal read_entry 3014590464 : seq 1524595 127 bytes -374> 2013-05-14 08:13:43.159537 7ff469167780 2 journal read_entry 3014594560 : seq 1524596 55 bytes -373> 2013-05-14 08:13:43.159554 7ff469167780 2 journal read_entry 3014598656 : seq 1524597 127 bytes -372> 2013-05-14 08:13:43.159562 7ff469167780 2 journal read_entry 3014602752 : seq 1524598 55 bytes -371> 2013-05-14 08:13:43.159577 7ff469167780 2 journal read_entry 3014606848 : seq 1524599 128 bytes -370> 2013-05-14 08:13:43.159588 7ff469167780 2 journal read_entry 3014610944 : seq 1524600 56 bytes -369> 2013-05-14 08:13:43.159597 7ff469167780 2 journal read_entry 3014615040 : seq 1524601 128 bytes -368> 2013-05-14 08:13:43.159606 7ff469167780 2 journal read_entry 3014619136 : seq 1524602 56 bytes -367> 2013-05-14 08:13:43.159615 7ff469167780 2 journal read_entry 3014623232 : seq 1524603 128 bytes -366> 2013-05-14 08:13:43.159624 7ff469167780 2 journal read_entry 3014627328 : seq 1524604 56 bytes -365> 2013-05-14 08:13:43.159634 7ff469167780 2 journal read_entry 3014631424 : seq 1524605 128 bytes -364> 2013-05-14 08:13:43.159643 7ff469167780 2 journal read_entry 3014635520 : seq 1524606 56 bytes -363> 2013-05-14 08:13:43.159652 7ff469167780 2 journal read_entry 3014639616 : seq 1524607 128 bytes -362> 2013-05-14 08:13:43.159661 7ff469167780 2 journal read_entry 3014643712 : seq 1524608 56 bytes -361> 2013-05-14 08:13:43.159671 7ff469167780 2 journal read_entry 3014647808 : seq 1524609 128 bytes -360> 2013-05-14 08:13:43.159680 7ff469167780 2 journal read_entry 3014651904 : seq 1524610 56 bytes -359> 2013-05-14 08:13:43.159704 7ff469167780 2 journal read_entry 3014660096 : seq 1524611 4398 bytes -358> 2013-05-14 08:13:43.159716 7ff469167780 2 journal read_entry 3014664192 : seq 1524612 798 bytes -357> 2013-05-14 08:13:43.160694 7ff469167780 2 journal read_entry 3014840320 : seq 1524613 172364 bytes -356> 2013-05-14 08:13:43.160714 7ff469167780 2 journal read_entry 3014844416 : seq 1524614 127 bytes -355> 2013-05-14 08:13:43.160724 7ff469167780 2 journal read_entry 3014848512 : seq 1524615 55 bytes -354> 2013-05-14 08:13:43.160737 7ff469167780 2 journal read_entry 3014852608 : seq 1524616 127 bytes -353> 2013-05-14 08:13:43.160747 7ff469167780 2 journal read_entry 3014856704 : seq 1524617 55 bytes -352> 2013-05-14 08:13:43.160757 7ff469167780 2 journal read_entry 3014860800 : seq 1524618 127 bytes -351> 2013-05-14 08:13:43.160766 7ff469167780 2 journal read_entry 3014864896 : seq 1524619 55 bytes -350> 2013-05-14 08:13:43.160775 7ff469167780 2 journal read_entry 3014868992 : seq 1524620 128 bytes -349> 2013-05-14 08:13:43.160784 7ff469167780 2 journal read_entry 3014873088 : seq 1524621 56 bytes -348> 2013-05-14 08:13:43.160793 7ff469167780 2 journal read_entry 3014877184 : seq 1524622 128 bytes -347> 2013-05-14 08:13:43.160802 7ff469167780 2 journal read_entry 3014881280 : seq 1524623 56 bytes -346> 2013-05-14 08:13:43.160812 7ff469167780 2 journal read_entry 3014885376 : seq 1524624 128 bytes -345> 2013-05-14 08:13:43.160822 7ff469167780 2 journal read_entry 3014889472 : seq 1524625 56 bytes -344> 2013-05-14 08:13:43.160835 7ff469167780 2 journal read_entry 3014893568 : seq 1524626 2548 bytes -343> 2013-05-14 08:13:43.160846 7ff469167780 2 journal read_entry 3014897664 : seq 1524627 2071 bytes -342> 2013-05-14 08:13:43.161875 7ff469167780 2 journal read_entry 3015102464 : seq 1524628 200708 bytes -341> 2013-05-14 08:13:43.161898 7ff469167780 2 journal read_entry 3015106560 : seq 1524629 215 bytes -340> 2013-05-14 08:13:43.161907 7ff469167780 2 journal read_entry 3015110656 : seq 1524630 55 bytes -339> 2013-05-14 08:13:43.161915 7ff469167780 2 journal read_entry 3015114752 : seq 1524631 215 bytes -338> 2013-05-14 08:13:43.161924 7ff469167780 2 journal read_entry 3015118848 : seq 1524632 55 bytes -337> 2013-05-14 08:13:43.161934 7ff469167780 2 journal read_entry 3015122944 : seq 1524633 128 bytes -336> 2013-05-14 08:13:43.161943 7ff469167780 2 journal read_entry 3015127040 : seq 1524634 56 bytes -335> 2013-05-14 08:13:43.162107 7ff469167780 2 journal read_entry 3015131136 : seq 1524635 128 bytes -334> 2013-05-14 08:13:43.162143 7ff469167780 2 journal read_entry 3015135232 : seq 1524636 56 bytes -333> 2013-05-14 08:13:43.162152 7ff469167780 2 journal read_entry 3015139328 : seq 1524637 128 bytes -332> 2013-05-14 08:13:43.162162 7ff469167780 2 journal read_entry 3015143424 : seq 1524638 56 bytes -331> 2013-05-14 08:13:43.162171 7ff469167780 2 journal read_entry 3015147520 : seq 1524639 128 bytes -330> 2013-05-14 08:13:43.162178 7ff469167780 2 journal read_entry 3015151616 : seq 1524640 56 bytes -329> 2013-05-14 08:13:43.162187 7ff469167780 2 journal read_entry 3015155712 : seq 1524641 128 bytes -328> 2013-05-14 08:13:43.162195 7ff469167780 2 journal read_entry 3015159808 : seq 1524642 56 bytes -327> 2013-05-14 08:13:43.162204 7ff469167780 2 journal read_entry 3015163904 : seq 1524643 128 bytes -326> 2013-05-14 08:13:43.162213 7ff469167780 2 journal read_entry 3015168000 : seq 1524644 56 bytes -325> 2013-05-14 08:13:43.162222 7ff469167780 2 journal read_entry 3015172096 : seq 1524645 128 bytes -324> 2013-05-14 08:13:43.162232 7ff469167780 2 journal read_entry 3015176192 : seq 1524646 56 bytes -323> 2013-05-14 08:13:43.162257 7ff469167780 2 journal read_entry 3015184384 : seq 1524647 5073 bytes -322> 2013-05-14 08:13:43.162272 7ff469167780 2 journal read_entry 3015188480 : seq 1524648 2283 bytes -321> 2013-05-14 08:13:43.162479 7ff469167780 2 journal read_entry 3015303168 : seq 1524649 111044 bytes -320> 2013-05-14 08:13:43.162495 7ff469167780 2 journal read_entry 3015307264 : seq 1524650 39 bytes -319> 2013-05-14 08:13:43.162540 7ff469167780 2 journal read_entry 3015331840 : seq 1524651 23465 bytes -318> 2013-05-14 08:13:43.163782 7ff469167780 2 journal read_entry 3015524352 : seq 1524652 192260 bytes -317> 2013-05-14 08:13:43.163802 7ff469167780 2 journal read_entry 3015528448 : seq 1524653 128 bytes -316> 2013-05-14 08:13:43.163813 7ff469167780 2 journal read_entry 3015532544 : seq 1524654 56 bytes -315> 2013-05-14 08:13:43.163822 7ff469167780 2 journal read_entry 3015536640 : seq 1524655 524 bytes -314> 2013-05-14 08:13:43.163830 7ff469167780 2 journal read_entry 3015540736 : seq 1524656 798 bytes -313> 2013-05-14 08:13:43.164952 7ff469167780 2 journal read_entry 3015737344 : seq 1524657 192464 bytes -312> 2013-05-14 08:13:43.164974 7ff469167780 2 journal read_entry 3015741440 : seq 1524658 127 bytes -311> 2013-05-14 08:13:43.164982 7ff469167780 2 journal read_entry 3015745536 : seq 1524659 55 bytes -310> 2013-05-14 08:13:43.164989 7ff469167780 2 journal read_entry 3015749632 : seq 1524660 127 bytes -309> 2013-05-14 08:13:43.164996 7ff469167780 2 journal read_entry 3015753728 : seq 1524661 55 bytes -308> 2013-05-14 08:13:43.165009 7ff469167780 2 journal read_entry 3015757824 : seq 1524662 128 bytes -307> 2013-05-14 08:13:43.165019 7ff469167780 2 journal read_entry 3015761920 : seq 1524663 56 bytes -306> 2013-05-14 08:13:43.165030 7ff469167780 2 journal read_entry 3015766016 : seq 1524664 128 bytes -305> 2013-05-14 08:13:43.165040 7ff469167780 2 journal read_entry 3015770112 : seq 1524665 56 bytes -304> 2013-05-14 08:13:43.165051 7ff469167780 2 journal read_entry 3015774208 : seq 1524666 128 bytes -303> 2013-05-14 08:13:43.165061 7ff469167780 2 journal read_entry 3015778304 : seq 1524667 56 bytes -302> 2013-05-14 08:13:43.165070 7ff469167780 2 journal read_entry 3015782400 : seq 1524668 128 bytes -301> 2013-05-14 08:13:43.165085 7ff469167780 2 journal read_entry 3015786496 : seq 1524669 56 bytes -300> 2013-05-14 08:13:43.165106 7ff469167780 2 journal read_entry 3015790592 : seq 1524670 3224 bytes -299> 2013-05-14 08:13:43.165120 7ff469167780 2 journal read_entry 3015794688 : seq 1524671 1887 bytes -298> 2013-05-14 08:13:43.165438 7ff469167780 2 journal read_entry 3015901184 : seq 1524672 106244 bytes -297> 2013-05-14 08:13:43.165459 7ff469167780 2 journal read_entry 3015905280 : seq 1524673 39 bytes -296> 2013-05-14 08:13:43.165468 7ff469167780 2 journal read_entry 3015909376 : seq 1524674 798 bytes -295> 2013-05-14 08:13:43.166231 7ff469167780 2 journal read_entry 3016101888 : seq 1524675 192260 bytes -294> 2013-05-14 08:13:43.166249 7ff469167780 2 journal read_entry 3016105984 : seq 1524676 128 bytes -293> 2013-05-14 08:13:43.166257 7ff469167780 2 journal read_entry 3016110080 : seq 1524677 56 bytes -292> 2013-05-14 08:13:43.166265 7ff469167780 2 journal read_entry 3016114176 : seq 1524678 524 bytes -291> 2013-05-14 08:13:43.166275 7ff469167780 2 journal read_entry 3016118272 : seq 1524679 2173 bytes -290> 2013-05-14 08:13:43.167383 7ff469167780 2 journal read_entry 3016298496 : seq 1524680 179216 bytes -289> 2013-05-14 08:13:43.167403 7ff469167780 2 journal read_entry 3016302592 : seq 1524681 128 bytes -288> 2013-05-14 08:13:43.167411 7ff469167780 2 journal read_entry 3016306688 : seq 1524682 56 bytes -287> 2013-05-14 08:13:43.167419 7ff469167780 2 journal read_entry 3016310784 : seq 1524683 524 bytes -286> 2013-05-14 08:13:43.167427 7ff469167780 2 journal read_entry 3016314880 : seq 1524684 798 bytes -285> 2013-05-14 08:13:43.167958 7ff469167780 2 journal read_entry 3016511488 : seq 1524685 192452 bytes -284> 2013-05-14 08:13:43.167977 7ff469167780 2 journal read_entry 3016515584 : seq 1524686 39 bytes -283> 2013-05-14 08:13:43.167987 7ff469167780 2 journal read_entry 3016519680 : seq 1524687 2079 bytes -282> 2013-05-14 08:13:43.168348 7ff469167780 2 journal No further valid entries found, journal is most likely valid -281> 2013-05-14 08:13:43.168362 7ff469167780 2 journal No further valid entries found, journal is most likely valid -280> 2013-05-14 08:13:43.168364 7ff469167780 3 journal journal_replay: end of journal, done. -279> 2013-05-14 08:13:43.171846 7ff469167780 1 journal _open /dev/sdi3 fd 18: 26843545600 bytes, block size 4096 bytes, directio = 1, aio = 1 -278> 2013-05-14 08:13:43.172533 7ff461e7e700 1 FileStore::op_tp worker finish -277> 2013-05-14 08:13:43.172626 7ff46167d700 1 FileStore::op_tp worker finish -276> 2013-05-14 08:13:43.172772 7ff469167780 1 journal close /dev/sdi3 -275> 2013-05-14 08:13:43.173616 7ff469167780 10 monclient(hunting): build_initial_monmap -274> 2013-05-14 08:13:43.173746 7ff469167780 5 adding auth protocol: cephx -273> 2013-05-14 08:13:43.173754 7ff469167780 5 adding auth protocol: cephx -272> 2013-05-14 08:13:43.173865 7ff469167780 1 -- 10.0.0.3:6806/31424 messenger.start -271> 2013-05-14 08:13:43.173899 7ff469167780 1 -- :/0 messenger.start -270> 2013-05-14 08:13:43.173916 7ff469167780 1 -- 192.168.42.3:6811/31424 messenger.start -269> 2013-05-14 08:13:43.173935 7ff469167780 1 -- 192.168.42.3:6810/31424 messenger.start -268> 2013-05-14 08:13:43.174066 7ff469167780 2 osd.25 0 mounting /var/lib/ceph/osd/ceph-25 /dev/sdi3 -267> 2013-05-14 08:13:43.179918 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount FIEMAP ioctl is supported and appears to work -266> 2013-05-14 08:13:43.179929 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount FIEMAP ioctl is disabled via 'filestore fiemap' config option -265> 2013-05-14 08:13:43.180340 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount did NOT detect btrfs -264> 2013-05-14 08:13:43.182242 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount syscall(__NR_syncfs, fd) fully supported -263> 2013-05-14 08:13:43.182304 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount found snaps <> -262> 2013-05-14 08:13:43.185621 7ff469167780 0 filestore(/var/lib/ceph/osd/ceph-25) mount: enabling WRITEAHEAD journal mode: btrfs not detected -261> 2013-05-14 08:13:43.190960 7ff469167780 2 journal open /dev/sdi3 fsid e82dd661-2ad1-4103-92d4-26a3926d524b fs_op_seq 1524687 -260> 2013-05-14 08:13:43.193318 7ff469167780 1 journal _open /dev/sdi3 fd 26: 26843545600 bytes, block size 4096 bytes, directio = 1, aio = 1 -259> 2013-05-14 08:13:43.195327 7ff469167780 2 journal read_entry 3012567040 : seq 1524509 111324 bytes -258> 2013-05-14 08:13:43.195410 7ff469167780 2 journal read_entry 3012571136 : seq 1524510 127 bytes -257> 2013-05-14 08:13:43.195428 7ff469167780 2 journal read_entry 3012575232 : seq 1524511 55 bytes -256> 2013-05-14 08:13:43.195436 7ff469167780 2 journal read_entry 3012579328 : seq 1524512 127 bytes -255> 2013-05-14 08:13:43.195443 7ff469167780 2 journal read_entry 3012583424 : seq 1524513 55 bytes -254> 2013-05-14 08:13:43.195450 7ff469167780 2 journal read_entry 3012587520 : seq 1524514 128 bytes -253> 2013-05-14 08:13:43.196063 7ff469167780 2 journal read_entry 3012591616 : seq 1524515 56 bytes -252> 2013-05-14 08:13:43.196078 7ff469167780 2 journal read_entry 3012595712 : seq 1524516 128 bytes -251> 2013-05-14 08:13:43.196085 7ff469167780 2 journal read_entry 3012599808 : seq 1524517 56 bytes -250> 2013-05-14 08:13:43.196092 7ff469167780 2 journal read_entry 3012603904 : seq 1524518 128 bytes -249> 2013-05-14 08:13:43.196099 7ff469167780 2 journal read_entry 3012608000 : seq 1524519 56 bytes -248> 2013-05-14 08:13:43.196106 7ff469167780 2 journal read_entry 3012612096 : seq 1524520 128 bytes -247> 2013-05-14 08:13:43.196119 7ff469167780 2 journal read_entry 3012616192 : seq 1524521 56 bytes -246> 2013-05-14 08:13:43.196128 7ff469167780 2 journal read_entry 3012620288 : seq 1524522 128 bytes -245> 2013-05-14 08:13:43.196140 7ff469167780 2 journal read_entry 3012624384 : seq 1524523 56 bytes -244> 2013-05-14 08:13:43.196150 7ff469167780 2 journal read_entry 3012628480 : seq 1524524 128 bytes -243> 2013-05-14 08:13:43.196159 7ff469167780 2 journal read_entry 3012632576 : seq 1524525 56 bytes -242> 2013-05-14 08:13:43.196179 7ff469167780 2 journal read_entry 3012636672 : seq 1524526 3327 bytes -241> 2013-05-14 08:13:43.196195 7ff469167780 2 journal read_entry 3012640768 : seq 1524527 2067 bytes -240> 2013-05-14 08:13:43.197046 7ff469167780 2 journal read_entry 3012833280 : seq 1524528 192260 bytes -239> 2013-05-14 08:13:43.197066 7ff469167780 2 journal read_entry 3012837376 : seq 1524529 39 bytes -238> 2013-05-14 08:13:43.197077 7ff469167780 2 journal read_entry 3012841472 : seq 1524530 798 bytes -237> 2013-05-14 08:13:43.198055 7ff469167780 2 journal read_entry 3013038080 : seq 1524531 194372 bytes -236> 2013-05-14 08:13:43.198076 7ff469167780 2 journal read_entry 3013042176 : seq 1524532 39 bytes -235> 2013-05-14 08:13:43.198085 7ff469167780 2 journal read_entry 3013046272 : seq 1524533 798 bytes -234> 2013-05-14 08:13:43.198532 7ff469167780 2 journal read_entry 3013201920 : seq 1524534 152708 bytes -233> 2013-05-14 08:13:43.198551 7ff469167780 2 journal read_entry 3013206016 : seq 1524535 39 bytes -232> 2013-05-14 08:13:43.198565 7ff469167780 2 journal read_entry 3013210112 : seq 1524536 2380 bytes -231> 2013-05-14 08:13:43.199227 7ff469167780 2 journal read_entry 3013369856 : seq 1524537 158184 bytes -230> 2013-05-14 08:13:43.199247 7ff469167780 2 journal read_entry 3013373952 : seq 1524538 127 bytes -229> 2013-05-14 08:13:43.199568 7ff469167780 2 journal read_entry 3013378048 : seq 1524539 55 bytes -228> 2013-05-14 08:13:43.199583 7ff469167780 2 journal read_entry 3013382144 : seq 1524540 127 bytes -227> 2013-05-14 08:13:43.199591 7ff469167780 2 journal read_entry 3013386240 : seq 1524541 55 bytes -226> 2013-05-14 08:13:43.199598 7ff469167780 2 journal read_entry 3013390336 : seq 1524542 128 bytes -225> 2013-05-14 08:13:43.199605 7ff469167780 2 journal read_entry 3013394432 : seq 1524543 56 bytes -224> 2013-05-14 08:13:43.199614 7ff469167780 2 journal read_entry 3013398528 : seq 1524544 128 bytes -223> 2013-05-14 08:13:43.199623 7ff469167780 2 journal read_entry 3013402624 : seq 1524545 56 bytes -222> 2013-05-14 08:13:43.199632 7ff469167780 2 journal read_entry 3013406720 : seq 1524546 128 bytes -221> 2013-05-14 08:13:43.199648 7ff469167780 2 journal read_entry 3013410816 : seq 1524547 56 bytes -220> 2013-05-14 08:13:43.199658 7ff469167780 2 journal read_entry 3013414912 : seq 1524548 128 bytes -219> 2013-05-14 08:13:43.199667 7ff469167780 2 journal read_entry 3013419008 : seq 1524549 56 bytes -218> 2013-05-14 08:13:43.199676 7ff469167780 2 journal read_entry 3013423104 : seq 1524550 128 bytes -217> 2013-05-14 08:13:43.199685 7ff469167780 2 journal read_entry 3013427200 : seq 1524551 56 bytes -216> 2013-05-14 08:13:43.199694 7ff469167780 2 journal read_entry 3013431296 : seq 1524552 128 bytes -215> 2013-05-14 08:13:43.199703 7ff469167780 2 journal read_entry 3013435392 : seq 1524553 56 bytes -214> 2013-05-14 08:13:43.199713 7ff469167780 2 journal read_entry 3013439488 : seq 1524554 128 bytes -213> 2013-05-14 08:13:43.199722 7ff469167780 2 journal read_entry 3013443584 : seq 1524555 56 bytes -212> 2013-05-14 08:13:43.199740 7ff469167780 2 journal read_entry 3013447680 : seq 1524556 3812 bytes -211> 2013-05-14 08:13:43.199756 7ff469167780 2 journal read_entry 3013451776 : seq 1524557 1969 bytes -210> 2013-05-14 08:13:43.200209 7ff469167780 2 journal read_entry 3013562368 : seq 1524558 107588 bytes -209> 2013-05-14 08:13:43.200227 7ff469167780 2 journal read_entry 3013566464 : seq 1524559 39 bytes -208> 2013-05-14 08:13:43.200241 7ff469167780 2 journal read_entry 3013570560 : seq 1524560 3641 bytes -207> 2013-05-14 08:13:43.200753 7ff469167780 2 journal read_entry 3013697536 : seq 1524561 126404 bytes -206> 2013-05-14 08:13:43.200771 7ff469167780 2 journal read_entry 3013701632 : seq 1524562 39 bytes -205> 2013-05-14 08:13:43.200781 7ff469167780 2 journal read_entry 3013705728 : seq 1524563 2004 bytes -204> 2013-05-14 08:13:43.201406 7ff469167780 2 journal read_entry 3013840896 : seq 1524564 134672 bytes -203> 2013-05-14 08:13:43.201426 7ff469167780 2 journal read_entry 3013844992 : seq 1524565 128 bytes -202> 2013-05-14 08:13:43.201439 7ff469167780 2 journal read_entry 3013849088 : seq 1524566 56 bytes -201> 2013-05-14 08:13:43.201447 7ff469167780 2 journal read_entry 3013853184 : seq 1524567 524 bytes -200> 2013-05-14 08:13:43.201456 7ff469167780 2 journal read_entry 3013857280 : seq 1524568 2145 bytes -199> 2013-05-14 08:13:43.202430 7ff469167780 2 journal read_entry 3014049792 : seq 1524569 192272 bytes -198> 2013-05-14 08:13:43.202450 7ff469167780 2 journal read_entry 3014053888 : seq 1524570 127 bytes -197> 2013-05-14 08:13:43.202457 7ff469167780 2 journal read_entry 3014057984 : seq 1524571 55 bytes -196> 2013-05-14 08:13:43.202465 7ff469167780 2 journal read_entry 3014062080 : seq 1524572 127 bytes -195> 2013-05-14 08:13:43.202472 7ff469167780 2 journal read_entry 3014066176 : seq 1524573 55 bytes -194> 2013-05-14 08:13:43.202480 7ff469167780 2 journal read_entry 3014070272 : seq 1524574 128 bytes -193> 2013-05-14 08:13:43.202487 7ff469167780 2 journal read_entry 3014074368 : seq 1524575 56 bytes -192> 2013-05-14 08:13:43.202494 7ff469167780 2 journal read_entry 3014078464 : seq 1524576 128 bytes -191> 2013-05-14 08:13:43.202503 7ff469167780 2 journal read_entry 3014082560 : seq 1524577 56 bytes -190> 2013-05-14 08:13:43.202512 7ff469167780 2 journal read_entry 3014086656 : seq 1524578 128 bytes -189> 2013-05-14 08:13:43.202522 7ff469167780 2 journal read_entry 3014090752 : seq 1524579 56 bytes -188> 2013-05-14 08:13:43.202529 7ff469167780 2 journal read_entry 3014094848 : seq 1524580 128 bytes -187> 2013-05-14 08:13:43.202538 7ff469167780 2 journal read_entry 3014098944 : seq 1524581 56 bytes -186> 2013-05-14 08:13:43.202547 7ff469167780 2 journal read_entry 3014103040 : seq 1524582 128 bytes -185> 2013-05-14 08:13:43.202556 7ff469167780 2 journal read_entry 3014107136 : seq 1524583 56 bytes -184> 2013-05-14 08:13:43.202576 7ff469167780 2 journal read_entry 3014111232 : seq 1524584 2930 bytes -183> 2013-05-14 08:13:43.202591 7ff469167780 2 journal read_entry 3014115328 : seq 1524585 798 bytes -182> 2013-05-14 08:13:43.203538 7ff469167780 2 journal read_entry 3014307840 : seq 1524586 192260 bytes -181> 2013-05-14 08:13:43.203558 7ff469167780 2 journal read_entry 3014311936 : seq 1524587 39 bytes -180> 2013-05-14 08:13:43.203567 7ff469167780 2 journal read_entry 3014316032 : seq 1524588 798 bytes -179> 2013-05-14 08:13:43.203643 7ff469167780 2 journal read_entry 3014373376 : seq 1524589 57092 bytes -178> 2013-05-14 08:13:43.203654 7ff469167780 2 journal read_entry 3014377472 : seq 1524590 39 bytes -177> 2013-05-14 08:13:43.203665 7ff469167780 2 journal read_entry 3014381568 : seq 1524591 2106 bytes -176> 2013-05-14 08:13:43.204727 7ff469167780 2 journal read_entry 3014578176 : seq 1524592 192452 bytes -175> 2013-05-14 08:13:43.204747 7ff469167780 2 journal read_entry 3014582272 : seq 1524593 127 bytes -174> 2013-05-14 08:13:43.204754 7ff469167780 2 journal read_entry 3014586368 : seq 1524594 55 bytes -173> 2013-05-14 08:13:43.204762 7ff469167780 2 journal read_entry 3014590464 : seq 1524595 127 bytes -172> 2013-05-14 08:13:43.204768 7ff469167780 2 journal read_entry 3014594560 : seq 1524596 55 bytes -171> 2013-05-14 08:13:43.204776 7ff469167780 2 journal read_entry 3014598656 : seq 1524597 127 bytes -170> 2013-05-14 08:13:43.204783 7ff469167780 2 journal read_entry 3014602752 : seq 1524598 55 bytes -169> 2013-05-14 08:13:43.204790 7ff469167780 2 journal read_entry 3014606848 : seq 1524599 128 bytes -168> 2013-05-14 08:13:43.204797 7ff469167780 2 journal read_entry 3014610944 : seq 1524600 56 bytes -167> 2013-05-14 08:13:43.204804 7ff469167780 2 journal read_entry 3014615040 : seq 1524601 128 bytes -166> 2013-05-14 08:13:43.204818 7ff469167780 2 journal read_entry 3014619136 : seq 1524602 56 bytes -165> 2013-05-14 08:13:43.204831 7ff469167780 2 journal read_entry 3014623232 : seq 1524603 128 bytes -164> 2013-05-14 08:13:43.204848 7ff469167780 2 journal read_entry 3014627328 : seq 1524604 56 bytes -163> 2013-05-14 08:13:43.204856 7ff469167780 2 journal read_entry 3014631424 : seq 1524605 128 bytes -162> 2013-05-14 08:13:43.204867 7ff469167780 2 journal read_entry 3014635520 : seq 1524606 56 bytes -161> 2013-05-14 08:13:43.204876 7ff469167780 2 journal read_entry 3014639616 : seq 1524607 128 bytes -160> 2013-05-14 08:13:43.204886 7ff469167780 2 journal read_entry 3014643712 : seq 1524608 56 bytes -159> 2013-05-14 08:13:43.204895 7ff469167780 2 journal read_entry 3014647808 : seq 1524609 128 bytes -158> 2013-05-14 08:13:43.204904 7ff469167780 2 journal read_entry 3014651904 : seq 1524610 56 bytes -157> 2013-05-14 08:13:43.204925 7ff469167780 2 journal read_entry 3014660096 : seq 1524611 4398 bytes -156> 2013-05-14 08:13:43.204936 7ff469167780 2 journal read_entry 3014664192 : seq 1524612 798 bytes -155> 2013-05-14 08:13:43.205919 7ff469167780 2 journal read_entry 3014840320 : seq 1524613 172364 bytes -154> 2013-05-14 08:13:43.205940 7ff469167780 2 journal read_entry 3014844416 : seq 1524614 127 bytes -153> 2013-05-14 08:13:43.205947 7ff469167780 2 journal read_entry 3014848512 : seq 1524615 55 bytes -152> 2013-05-14 08:13:43.205954 7ff469167780 2 journal read_entry 3014852608 : seq 1524616 127 bytes -151> 2013-05-14 08:13:43.205962 7ff469167780 2 journal read_entry 3014856704 : seq 1524617 55 bytes -150> 2013-05-14 08:13:43.205971 7ff469167780 2 journal read_entry 3014860800 : seq 1524618 127 bytes -149> 2013-05-14 08:13:43.205980 7ff469167780 2 journal read_entry 3014864896 : seq 1524619 55 bytes -148> 2013-05-14 08:13:43.205990 7ff469167780 2 journal read_entry 3014868992 : seq 1524620 128 bytes -147> 2013-05-14 08:13:43.205999 7ff469167780 2 journal read_entry 3014873088 : seq 1524621 56 bytes -146> 2013-05-14 08:13:43.206006 7ff469167780 2 journal read_entry 3014877184 : seq 1524622 128 bytes -145> 2013-05-14 08:13:43.206018 7ff469167780 2 journal read_entry 3014881280 : seq 1524623 56 bytes -144> 2013-05-14 08:13:43.206025 7ff469167780 2 journal read_entry 3014885376 : seq 1524624 128 bytes -143> 2013-05-14 08:13:43.206034 7ff469167780 2 journal read_entry 3014889472 : seq 1524625 56 bytes -142> 2013-05-14 08:13:43.206044 7ff469167780 2 journal read_entry 3014893568 : seq 1524626 2548 bytes -141> 2013-05-14 08:13:43.206056 7ff469167780 2 journal read_entry 3014897664 : seq 1524627 2071 bytes -140> 2013-05-14 08:13:43.207021 7ff469167780 2 journal read_entry 3015102464 : seq 1524628 200708 bytes -139> 2013-05-14 08:13:43.207044 7ff469167780 2 journal read_entry 3015106560 : seq 1524629 215 bytes -138> 2013-05-14 08:13:43.207052 7ff469167780 2 journal read_entry 3015110656 : seq 1524630 55 bytes -137> 2013-05-14 08:13:43.207059 7ff469167780 2 journal read_entry 3015114752 : seq 1524631 215 bytes -136> 2013-05-14 08:13:43.207066 7ff469167780 2 journal read_entry 3015118848 : seq 1524632 55 bytes -135> 2013-05-14 08:13:43.207076 7ff469167780 2 journal read_entry 3015122944 : seq 1524633 128 bytes -134> 2013-05-14 08:13:43.207087 7ff469167780 2 journal read_entry 3015127040 : seq 1524634 56 bytes -133> 2013-05-14 08:13:43.207097 7ff469167780 2 journal read_entry 3015131136 : seq 1524635 128 bytes -132> 2013-05-14 08:13:43.207106 7ff469167780 2 journal read_entry 3015135232 : seq 1524636 56 bytes -131> 2013-05-14 08:13:43.207114 7ff469167780 2 journal read_entry 3015139328 : seq 1524637 128 bytes -130> 2013-05-14 08:13:43.207122 7ff469167780 2 journal read_entry 3015143424 : seq 1524638 56 bytes -129> 2013-05-14 08:13:43.207130 7ff469167780 2 journal read_entry 3015147520 : seq 1524639 128 bytes -128> 2013-05-14 08:13:43.207139 7ff469167780 2 journal read_entry 3015151616 : seq 1524640 56 bytes -127> 2013-05-14 08:13:43.207146 7ff469167780 2 journal read_entry 3015155712 : seq 1524641 128 bytes -126> 2013-05-14 08:13:43.207158 7ff469167780 2 journal read_entry 3015159808 : seq 1524642 56 bytes -125> 2013-05-14 08:13:43.207165 7ff469167780 2 journal read_entry 3015163904 : seq 1524643 128 bytes -124> 2013-05-14 08:13:43.207174 7ff469167780 2 journal read_entry 3015168000 : seq 1524644 56 bytes -123> 2013-05-14 08:13:43.207181 7ff469167780 2 journal read_entry 3015172096 : seq 1524645 128 bytes -122> 2013-05-14 08:13:43.207190 7ff469167780 2 journal read_entry 3015176192 : seq 1524646 56 bytes -121> 2013-05-14 08:13:43.207211 7ff469167780 2 journal read_entry 3015184384 : seq 1524647 5073 bytes -120> 2013-05-14 08:13:43.207225 7ff469167780 2 journal read_entry 3015188480 : seq 1524648 2283 bytes -119> 2013-05-14 08:13:43.207451 7ff469167780 2 journal read_entry 3015303168 : seq 1524649 111044 bytes -118> 2013-05-14 08:13:43.207467 7ff469167780 2 journal read_entry 3015307264 : seq 1524650 39 bytes -117> 2013-05-14 08:13:43.207520 7ff469167780 2 journal read_entry 3015331840 : seq 1524651 23465 bytes -116> 2013-05-14 08:13:43.208733 7ff469167780 2 journal read_entry 3015524352 : seq 1524652 192260 bytes -115> 2013-05-14 08:13:43.208753 7ff469167780 2 journal read_entry 3015528448 : seq 1524653 128 bytes -114> 2013-05-14 08:13:43.208763 7ff469167780 2 journal read_entry 3015532544 : seq 1524654 56 bytes -113> 2013-05-14 08:13:43.208771 7ff469167780 2 journal read_entry 3015536640 : seq 1524655 524 bytes -112> 2013-05-14 08:13:43.208779 7ff469167780 2 journal read_entry 3015540736 : seq 1524656 798 bytes -111> 2013-05-14 08:13:43.209724 7ff469167780 2 journal read_entry 3015737344 : seq 1524657 192464 bytes -110> 2013-05-14 08:13:43.209744 7ff469167780 2 journal read_entry 3015741440 : seq 1524658 127 bytes -109> 2013-05-14 08:13:43.209752 7ff469167780 2 journal read_entry 3015745536 : seq 1524659 55 bytes -108> 2013-05-14 08:13:43.209759 7ff469167780 2 journal read_entry 3015749632 : seq 1524660 127 bytes -107> 2013-05-14 08:13:43.209769 7ff469167780 2 journal read_entry 3015753728 : seq 1524661 55 bytes -106> 2013-05-14 08:13:43.209780 7ff469167780 2 journal read_entry 3015757824 : seq 1524662 128 bytes -105> 2013-05-14 08:13:43.209792 7ff469167780 2 journal read_entry 3015761920 : seq 1524663 56 bytes -104> 2013-05-14 08:13:43.209802 7ff469167780 2 journal read_entry 3015766016 : seq 1524664 128 bytes -103> 2013-05-14 08:13:43.209812 7ff469167780 2 journal read_entry 3015770112 : seq 1524665 56 bytes -102> 2013-05-14 08:13:43.209821 7ff469167780 2 journal read_entry 3015774208 : seq 1524666 128 bytes -101> 2013-05-14 08:13:43.209828 7ff469167780 2 journal read_entry 3015778304 : seq 1524667 56 bytes -100> 2013-05-14 08:13:43.209835 7ff469167780 2 journal read_entry 3015782400 : seq 1524668 128 bytes -99> 2013-05-14 08:13:43.209850 7ff469167780 2 journal read_entry 3015786496 : seq 1524669 56 bytes -98> 2013-05-14 08:13:43.209864 7ff469167780 2 journal read_entry 3015790592 : seq 1524670 3224 bytes -97> 2013-05-14 08:13:43.209875 7ff469167780 2 journal read_entry 3015794688 : seq 1524671 1887 bytes -96> 2013-05-14 08:13:43.210260 7ff469167780 2 journal read_entry 3015901184 : seq 1524672 106244 bytes -95> 2013-05-14 08:13:43.210278 7ff469167780 2 journal read_entry 3015905280 : seq 1524673 39 bytes -94> 2013-05-14 08:13:43.210286 7ff469167780 2 journal read_entry 3015909376 : seq 1524674 798 bytes -93> 2013-05-14 08:13:43.210838 7ff469167780 2 journal read_entry 3016101888 : seq 1524675 192260 bytes -92> 2013-05-14 08:13:43.210856 7ff469167780 2 journal read_entry 3016105984 : seq 1524676 128 bytes -91> 2013-05-14 08:13:43.210865 7ff469167780 2 journal read_entry 3016110080 : seq 1524677 56 bytes -90> 2013-05-14 08:13:43.210873 7ff469167780 2 journal read_entry 3016114176 : seq 1524678 524 bytes -89> 2013-05-14 08:13:43.210883 7ff469167780 2 journal read_entry 3016118272 : seq 1524679 2173 bytes -88> 2013-05-14 08:13:43.211899 7ff469167780 2 journal read_entry 3016298496 : seq 1524680 179216 bytes -87> 2013-05-14 08:13:43.211919 7ff469167780 2 journal read_entry 3016302592 : seq 1524681 128 bytes -86> 2013-05-14 08:13:43.211926 7ff469167780 2 journal read_entry 3016306688 : seq 1524682 56 bytes -85> 2013-05-14 08:13:43.211934 7ff469167780 2 journal read_entry 3016310784 : seq 1524683 524 bytes -84> 2013-05-14 08:13:43.211943 7ff469167780 2 journal read_entry 3016314880 : seq 1524684 798 bytes -83> 2013-05-14 08:13:43.212475 7ff469167780 2 journal read_entry 3016511488 : seq 1524685 192452 bytes -82> 2013-05-14 08:13:43.212494 7ff469167780 2 journal read_entry 3016515584 : seq 1524686 39 bytes -81> 2013-05-14 08:13:43.212504 7ff469167780 2 journal read_entry 3016519680 : seq 1524687 2079 bytes -80> 2013-05-14 08:13:43.212810 7ff469167780 2 journal No further valid entries found, journal is most likely valid -79> 2013-05-14 08:13:43.212829 7ff469167780 2 journal No further valid entries found, journal is most likely valid -78> 2013-05-14 08:13:43.212831 7ff469167780 3 journal journal_replay: end of journal, done. -77> 2013-05-14 08:13:43.216180 7ff469167780 1 journal _open /dev/sdi3 fd 26: 26843545600 bytes, block size 4096 bytes, directio = 1, aio = 1 -76> 2013-05-14 08:13:43.216647 7ff469167780 2 osd.25 0 boot -75> 2013-05-14 08:13:43.253760 7ff469167780 0 osd.25 58876 crush map has features 262144, adjusting msgr requires for clients -74> 2013-05-14 08:13:43.253776 7ff469167780 0 osd.25 58876 crush map has features 262144, adjusting msgr requires for osds -73> 2013-05-14 08:13:48.569764 7ff469167780 2 osd.25 58876 superblock: i am osd.25 -72> 2013-05-14 08:13:48.569815 7ff469167780 1 accepter.accepter.start -71> 2013-05-14 08:13:48.569854 7ff469167780 1 accepter.accepter.start -70> 2013-05-14 08:13:48.569965 7ff469167780 1 accepter.accepter.start -69> 2013-05-14 08:13:48.569985 7ff469167780 10 monclient(hunting): init -68> 2013-05-14 08:13:48.569992 7ff469167780 5 adding auth protocol: cephx -67> 2013-05-14 08:13:48.569993 7ff469167780 10 monclient(hunting): auth_supported 2 method cephx -66> 2013-05-14 08:13:48.570143 7ff469167780 2 auth: KeyRing::load: loaded key file /etc/ceph/ceph.osd.25.keyring -65> 2013-05-14 08:13:48.570489 7ff469167780 5 asok(0x1fbe000) register_command dump_ops_in_flight hook 0xb074b40 -64> 2013-05-14 08:13:48.570497 7ff469167780 5 asok(0x1fbe000) register_command dump_historic_ops hook 0xb074b40 -63> 2013-05-14 08:13:48.570500 7ff469167780 5 asok(0x1fbe000) register_command dump_op_pq_state hook 0xb074b40 -62> 2013-05-14 08:13:48.570505 7ff469167780 5 asok(0x1fbe000) register_command setomapval hook 0xac9fd40 -61> 2013-05-14 08:13:48.570507 7ff469167780 5 asok(0x1fbe000) register_command rmomapkey hook 0xac9fd40 -60> 2013-05-14 08:13:48.570509 7ff469167780 5 asok(0x1fbe000) register_command setomapheader hook 0xac9fd40 -59> 2013-05-14 08:13:48.570512 7ff469167780 5 asok(0x1fbe000) register_command getomap hook 0xac9fd40 -58> 2013-05-14 08:13:48.570514 7ff469167780 5 asok(0x1fbe000) register_command truncobj hook 0xac9fd40 -57> 2013-05-14 08:13:48.570516 7ff469167780 5 asok(0x1fbe000) register_command injectdataerr hook 0xac9fd40 -56> 2013-05-14 08:13:48.570517 7ff469167780 5 asok(0x1fbe000) register_command injectmdataerr hook 0xac9fd40 -55> 2013-05-14 08:13:48.570573 7ff44b86c700 5 osd.25 58876 heartbeat: osd_stat(653 GB used, 272 GB avail, 926 GB total, peers []/[]) -54> 2013-05-14 08:13:48.571104 7ff469167780 10 monclient(hunting): _reopen_session -53> 2013-05-14 08:13:48.571177 7ff469167780 10 monclient(hunting): _pick_new_mon picked mon.e con 0x2024c60 addr 10.0.0.3:6789/0 -52> 2013-05-14 08:13:48.571201 7ff469167780 10 monclient(hunting): _send_mon_message to mon.e at 10.0.0.3:6789/0 -51> 2013-05-14 08:13:48.571212 7ff469167780 1 -- 10.0.0.3:6806/31424 --> 10.0.0.3:6789/0 -- auth(proto 0 27 bytes epoch 0) v1 -- ?+0 0x1fbdd80 con 0x2024c60 -50> 2013-05-14 08:13:48.571231 7ff469167780 10 monclient(hunting): renew_subs -49> 2013-05-14 08:13:48.571933 7ff457083700 10 monclient(hunting): renew_subs -48> 2013-05-14 08:13:51.570308 7ff45387c700 10 monclient(hunting): tick -47> 2013-05-14 08:13:51.570363 7ff45387c700 1 monclient(hunting): continuing hunt -46> 2013-05-14 08:13:51.570366 7ff45387c700 10 monclient(hunting): _reopen_session -45> 2013-05-14 08:13:51.570372 7ff45387c700 1 -- 10.0.0.3:6806/31424 mark_down 0x2024c60 -- 0xb2c7280 -44> 2013-05-14 08:13:51.570554 7ff449f68700 2 -- 10.0.0.3:6806/31424 >> 10.0.0.3:6789/0 pipe(0xb2c7280 sd=27 :8988 s=4 pgs=137817 cs=1 l=1).reader couldn't read tag, Success -43> 2013-05-14 08:13:51.570610 7ff449f68700 2 -- 10.0.0.3:6806/31424 >> 10.0.0.3:6789/0 pipe(0xb2c7280 sd=27 :8988 s=4 pgs=137817 cs=1 l=1).fault 0: Success -42> 2013-05-14 08:13:51.570623 7ff45387c700 10 monclient(hunting): _pick_new_mon picked mon.b con 0x2024dc0 addr 10.0.0.2:6789/0 -41> 2013-05-14 08:13:51.570644 7ff45387c700 10 monclient(hunting): _send_mon_message to mon.b at 10.0.0.2:6789/0 -40> 2013-05-14 08:13:51.570648 7ff45387c700 1 -- 10.0.0.3:6806/31424 --> 10.0.0.2:6789/0 -- auth(proto 0 27 bytes epoch 0) v1 -- ?+0 0xb2c3b40 con 0x2024dc0 -39> 2013-05-14 08:13:51.570657 7ff45387c700 10 monclient(hunting): renew_subs -38> 2013-05-14 08:13:51.571471 7ff457083700 10 monclient(hunting): renew_subs -37> 2013-05-14 08:13:51.657613 7ff457083700 1 -- 10.0.0.3:6806/31424 <== mon.1 10.0.0.2:6789/0 1 ==== mon_map v1 ==== 755+0+0 (1964479241 0 0) 0xb2cb000 con 0x2024dc0 -36> 2013-05-14 08:13:51.657694 7ff457083700 10 monclient(hunting): handle_monmap mon_map v1 -35> 2013-05-14 08:13:51.657726 7ff457083700 10 monclient(hunting): got monmap 7, mon.b is now rank 1 -34> 2013-05-14 08:13:51.657732 7ff457083700 10 monclient(hunting): dump: epoch 7 fsid de035250-323d-4cf6-8c4b-cf0faf6296b1 last_changed 2013-03-27 19:40:05.413825 created 2012-12-10 11:44:55.644633 0: 10.0.0.1:6789/0 mon.a 1: 10.0.0.2:6789/0 mon.b 2: 10.0.0.3:6789/0 mon.e 3: 10.0.0.5:6789/0 mon.c 4: 10.0.0.6:6789/0 mon.d -33> 2013-05-14 08:13:51.657777 7ff457083700 1 monclient(hunting): found mon.b -32> 2013-05-14 08:13:51.657836 7ff457083700 1 -- 10.0.0.3:6806/31424 <== mon.1 10.0.0.2:6789/0 2 ==== auth_reply(proto 2 0 Success) v1 ==== 33+0+0 (2634832749 0 0) 0xb2cb400 con 0x2024dc0 -31> 2013-05-14 08:13:51.657903 7ff457083700 10 monclient: my global_id is 12258331 -30> 2013-05-14 08:13:51.658092 7ff457083700 10 monclient: _send_mon_message to mon.b at 10.0.0.2:6789/0 -29> 2013-05-14 08:13:51.658105 7ff457083700 1 -- 10.0.0.3:6806/31424 --> 10.0.0.2:6789/0 -- auth(proto 2 32 bytes epoch 0) v1 -- ?+0 0x1fbdd80 con 0x2024dc0 -28> 2013-05-14 08:13:51.658953 7ff457083700 1 -- 10.0.0.3:6806/31424 <== mon.1 10.0.0.2:6789/0 3 ==== auth_reply(proto 2 0 Success) v1 ==== 206+0+0 (2176813729 0 0) 0xb2cb200 con 0x2024dc0 -27> 2013-05-14 08:13:51.659076 7ff457083700 10 monclient: _send_mon_message to mon.b at 10.0.0.2:6789/0 -26> 2013-05-14 08:13:51.659085 7ff457083700 1 -- 10.0.0.3:6806/31424 --> 10.0.0.2:6789/0 -- auth(proto 2 165 bytes epoch 0) v1 -- ?+0 0xb2c46c0 con 0x2024dc0 -25> 2013-05-14 08:13:51.659797 7ff457083700 1 -- 10.0.0.3:6806/31424 <== mon.1 10.0.0.2:6789/0 4 ==== auth_reply(proto 2 0 Success) v1 ==== 393+0+0 (1806667289 0 0) 0xb2cb600 con 0x2024dc0 -24> 2013-05-14 08:13:51.659864 7ff457083700 10 monclient: _send_mon_message to mon.b at 10.0.0.2:6789/0 -23> 2013-05-14 08:13:51.659870 7ff457083700 1 -- 10.0.0.3:6806/31424 --> 10.0.0.2:6789/0 -- mon_subscribe({monmap=0+,osd_pg_creates=0}) v2 -- ?+0 0x1febe00 con 0x2024dc0 -22> 2013-05-14 08:13:51.659890 7ff457083700 10 monclient: _send_mon_message to mon.b at 10.0.0.2:6789/0 -21> 2013-05-14 08:13:51.659893 7ff457083700 1 -- 10.0.0.3:6806/31424 --> 10.0.0.2:6789/0 -- mon_subscribe({monmap=0+,osd_pg_creates=0}) v2 -- ?+0 0xb2c81c0 con 0x2024dc0 -20> 2013-05-14 08:13:51.659909 7ff457083700 10 monclient: _check_auth_rotating renewing rotating keys (they expired before 2013-05-14 08:13:21.659907) -19> 2013-05-14 08:13:51.659917 7ff457083700 10 monclient: _send_mon_message to mon.b at 10.0.0.2:6789/0 -18> 2013-05-14 08:13:51.659919 7ff457083700 1 -- 10.0.0.3:6806/31424 --> 10.0.0.2:6789/0 -- auth(proto 2 2 bytes epoch 0) v1 -- ?+0 0xb2c3d80 con 0x2024dc0 -17> 2013-05-14 08:13:51.660022 7ff469167780 5 monclient: authenticate success, global_id 12258331 -16> 2013-05-14 08:13:51.660040 7ff469167780 10 monclient: wait_auth_rotating waiting (until 2013-05-14 08:14:21.660038) -15> 2013-05-14 08:13:51.660209 7ff457083700 1 -- 10.0.0.3:6806/31424 <== mon.1 10.0.0.2:6789/0 5 ==== mon_map v1 ==== 755+0+0 (1964479241 0 0) 0xb2cb400 con 0x2024dc0 -14> 2013-05-14 08:13:51.660228 7ff457083700 10 monclient: handle_monmap mon_map v1 -13> 2013-05-14 08:13:51.660242 7ff457083700 10 monclient: got monmap 7, mon.b is now rank 1 -12> 2013-05-14 08:13:51.660246 7ff457083700 10 monclient: dump: epoch 7 fsid de035250-323d-4cf6-8c4b-cf0faf6296b1 last_changed 2013-03-27 19:40:05.413825 created 2012-12-10 11:44:55.644633 0: 10.0.0.1:6789/0 mon.a 1: 10.0.0.2:6789/0 mon.b 2: 10.0.0.3:6789/0 mon.e 3: 10.0.0.5:6789/0 mon.c 4: 10.0.0.6:6789/0 mon.d -11> 2013-05-14 08:13:51.660298 7ff457083700 1 -- 10.0.0.3:6806/31424 <== mon.1 10.0.0.2:6789/0 6 ==== mon_subscribe_ack(300s) v1 ==== 20+0+0 (2197888173 0 0) 0xb2c81c0 con 0x2024dc0 -10> 2013-05-14 08:13:51.660309 7ff457083700 10 monclient: handle_subscribe_ack sent 2013-05-14 08:13:48.571239 renew after 2013-05-14 08:16:18.571239 -9> 2013-05-14 08:13:51.660338 7ff457083700 1 -- 10.0.0.3:6806/31424 <== mon.1 10.0.0.2:6789/0 7 ==== mon_map v1 ==== 755+0+0 (1964479241 0 0) 0xb2cb000 con 0x2024dc0 -8> 2013-05-14 08:13:51.660342 7ff457083700 10 monclient: handle_monmap mon_map v1 -7> 2013-05-14 08:13:51.660349 7ff457083700 10 monclient: got monmap 7, mon.b is now rank 1 -6> 2013-05-14 08:13:51.660353 7ff457083700 10 monclient: dump: epoch 7 fsid de035250-323d-4cf6-8c4b-cf0faf6296b1 last_changed 2013-03-27 19:40:05.413825 created 2012-12-10 11:44:55.644633 0: 10.0.0.1:6789/0 mon.a 1: 10.0.0.2:6789/0 mon.b 2: 10.0.0.3:6789/0 mon.e 3: 10.0.0.5:6789/0 mon.c 4: 10.0.0.6:6789/0 mon.d -5> 2013-05-14 08:13:51.660370 7ff457083700 1 -- 10.0.0.3:6806/31424 <== mon.1 10.0.0.2:6789/0 8 ==== mon_subscribe_ack(300s) v1 ==== 20+0+0 (2197888173 0 0) 0x1febe00 con 0x2024dc0 -4> 2013-05-14 08:13:51.660376 7ff457083700 10 monclient: handle_subscribe_ack sent 0.000000, ignoring -3> 2013-05-14 08:13:51.660405 7ff457083700 1 -- 10.0.0.3:6806/31424 <== mon.1 10.0.0.2:6789/0 9 ==== auth_reply(proto 2 0 Success) v1 ==== 194+0+0 (2034468970 0 0) 0xb2cb800 con 0x2024dc0 -2> 2013-05-14 08:13:51.660460 7ff457083700 10 monclient: _check_auth_rotating have uptodate secrets (they expire after 2013-05-14 08:13:21.660459) -1> 2013-05-14 08:13:51.660583 7ff469167780 10 monclient: wait_auth_rotating done 0> 2013-05-14 08:13:51.675152 7ff450075700 -1 osd/OSD.cc: In function 'OSDMapRef OSDService::get_map(epoch_t)' thread 7ff450075700 time 2013-05-14 08:13:51.661081 osd/OSD.cc: 4838: FAILED assert(_get_map_bl(epoch, bl)) ceph version 0.61-11-g3b94f03 (3b94f03ec58abe3d7a6d0359ff9b4d75826f3777) 1: (OSDService::get_map(unsigned int)+0x918) [0x608268] 2: (OSD::advance_pg(unsigned int, PG*, ThreadPool::TPHandle&, PG::RecoveryCtx*, std::set<boost::intrusive_ptr<PG>, std::less<boost::intrusive_ptr<PG> >, std::allocator<boost::intrusive_ptr<PG> > >*)+0x195) [0x60b725] 3: (OSD::process_peering_events(std::list<PG*, std::allocator<PG*> > const&, ThreadPool::TPHandle&)+0x24b) [0x630c1b] 4: (OSD::PeeringWQ::_process(std::list<PG*, std::allocator<PG*> > const&, ThreadPool::TPHandle&)+0x16) [0x6650b6] 5: (ThreadPool::worker(ThreadPool::WorkThread*)+0x561) [0x813a11] 6: (ThreadPool::WorkThread::entry()+0x10) [0x816530] 7: (()+0x68ca) [0x7ff4686028ca] 8: (clone()+0x6d) [0x7ff466ee0b6d] NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this. --- logging levels --- 0/ 5 none 0/ 1 lockdep 0/ 1 context 1/ 1 crush 1/ 5 mds 1/ 5 mds_balancer 1/ 5 mds_locker 1/ 5 mds_log 1/ 5 mds_log_expire 1/ 5 mds_migrator 0/ 1 buffer 0/ 1 timer 0/ 1 filer 0/ 1 striper 0/ 1 objecter 0/ 5 rados 0/ 5 rbd 0/ 5 journaler 0/ 5 objectcacher 0/ 5 client 0/ 5 osd 0/ 5 optracker 0/ 5 objclass 1/ 3 filestore 1/ 3 journal 0/ 5 ms 1/ 5 mon 0/10 monc 0/ 5 paxos 0/ 5 tp 1/ 5 auth 1/ 5 crypto 1/ 1 finisher 1/ 5 heartbeatmap 1/ 5 perfcounter 1/ 5 rgw 1/ 5 hadoop 1/ 5 javaclient 1/ 5 asok 1/ 1 throttle -1/-1 (syslog threshold) -1/-1 (stderr threshold) max_recent 10000 max_new 1000 log_file /var/log/ceph/osd.25.log --- end dump of recent events --- 2013-05-14 08:13:51.685837 7ff450075700 -1 *** Caught signal (Aborted) ** in thread 7ff450075700 ceph version 0.61-11-g3b94f03 (3b94f03ec58abe3d7a6d0359ff9b4d75826f3777) 1: /usr/bin/ceph-osd() [0x7a14f9] 2: (()+0xeff0) [0x7ff46860aff0] 3: (gsignal()+0x35) [0x7ff466e431b5] 4: (abort()+0x180) [0x7ff466e45fc0] 5: (__gnu_cxx::__verbose_terminate_handler()+0x115) [0x7ff4676d7dc5] 6: (()+0xcb166) [0x7ff4676d6166] 7: (()+0xcb193) [0x7ff4676d6193] 8: (()+0xcb28e) [0x7ff4676d628e] 9: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x7c9) [0x902c89] 10: (OSDService::get_map(unsigned int)+0x918) [0x608268] 11: (OSD::advance_pg(unsigned int, PG*, ThreadPool::TPHandle&, PG::RecoveryCtx*, std::set<boost::intrusive_ptr<PG>, std::less<boost::intrusive_ptr<PG> >, std::allocator<boost::intrusive_ptr<PG> > >*)+0x195) [0x60b725] 12: (OSD::process_peering_events(std::list<PG*, std::allocator<PG*> > const&, ThreadPool::TPHandle&)+0x24b) [0x630c1b] 13: (OSD::PeeringWQ::_process(std::list<PG*, std::allocator<PG*> > const&, ThreadPool::TPHandle&)+0x16) [0x6650b6] 14: (ThreadPool::worker(ThreadPool::WorkThread*)+0x561) [0x813a11] 15: (ThreadPool::WorkThread::entry()+0x10) [0x816530] 16: (()+0x68ca) [0x7ff4686028ca] 17: (clone()+0x6d) [0x7ff466ee0b6d] NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this. --- begin dump of recent events --- 0> 2013-05-14 08:13:51.685837 7ff450075700 -1 *** Caught signal (Aborted) ** in thread 7ff450075700 ceph version 0.61-11-g3b94f03 (3b94f03ec58abe3d7a6d0359ff9b4d75826f3777) 1: /usr/bin/ceph-osd() [0x7a14f9] 2: (()+0xeff0) [0x7ff46860aff0] 3: (gsignal()+0x35) [0x7ff466e431b5] 4: (abort()+0x180) [0x7ff466e45fc0] 5: (__gnu_cxx::__verbose_terminate_handler()+0x115) [0x7ff4676d7dc5] 6: (()+0xcb166) [0x7ff4676d6166] 7: (()+0xcb193) [0x7ff4676d6193] 8: (()+0xcb28e) [0x7ff4676d628e] 9: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x7c9) [0x902c89] 10: (OSDService::get_map(unsigned int)+0x918) [0x608268] 11: (OSD::advance_pg(unsigned int, PG*, ThreadPool::TPHandle&, PG::RecoveryCtx*, std::set<boost::intrusive_ptr<PG>, std::less<boost::intrusive_ptr<PG> >, std::allocator<boost::intrusive_ptr<PG> > >*)+0x195) [0x60b725] 12: (OSD::process_peering_events(std::list<PG*, std::allocator<PG*> > const&, ThreadPool::TPHandle&)+0x24b) [0x630c1b] 13: (OSD::PeeringWQ::_process(std::list<PG*, std::allocator<PG*> > const&, ThreadPool::TPHandle&)+0x16) [0x6650b6] 14: (ThreadPool::worker(ThreadPool::WorkThread*)+0x561) [0x813a11] 15: (ThreadPool::WorkThread::entry()+0x10) [0x816530] 16: (()+0x68ca) [0x7ff4686028ca] 17: (clone()+0x6d) [0x7ff466ee0b6d] NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this. --- logging levels --- 0/ 5 none 0/ 1 lockdep 0/ 1 context 1/ 1 crush 1/ 5 mds 1/ 5 mds_balancer 1/ 5 mds_locker 1/ 5 mds_log 1/ 5 mds_log_expire 1/ 5 mds_migrator 0/ 1 buffer 0/ 1 timer 0/ 1 filer 0/ 1 striper 0/ 1 objecter 0/ 5 rados 0/ 5 rbd 0/ 5 journaler 0/ 5 objectcacher 0/ 5 client 0/ 5 osd 0/ 5 optracker 0/ 5 objclass 1/ 3 filestore 1/ 3 journal 0/ 5 ms 1/ 5 mon 0/10 monc 0/ 5 paxos 0/ 5 tp 1/ 5 auth 1/ 5 crypto 1/ 1 finisher 1/ 5 heartbeatmap 1/ 5 perfcounter 1/ 5 rgw 1/ 5 hadoop 1/ 5 javaclient 1/ 5 asok 1/ 1 throttle -1/-1 (syslog threshold) -1/-1 (stderr threshold) max_recent 10000 max_new 1000 log_file /var/log/ceph/osd.25.log --- end dump of recent events --- Le vendredi 17 mai 2013 à 11:36 -0700, John Wilkins a écrit : > Another thing... since your osd.10 is near full, your cluster may be > fairly close to capacity for the purposes of rebalancing. Have a look > at: > > http://ceph.com/docs/master/rados/configuration/mon-config-ref/#storage-capacity > http://ceph.com/docs/master/rados/troubleshooting/troubleshooting-osd/#no-free-drive-space > > Maybe we can get some others to look at this. It's not clear to me > why the other OSD crashes after you take osd.25 out. It could be > capacity, but that shouldn't make it crash. Have you tried adding more > OSDs to increase capacity? > > > > On Fri, May 17, 2013 at 11:27 AM, John Wilkins <john.wilkins@xxxxxxxxxxx> wrote: > > It looks like you have the "noout" flag set: > > > > "noout flag(s) set; 1 mons down, quorum 0,1,2,3 a,b,c,e > > monmap e7: 5 mons at > > {a=10.0.0.1:6789/0,b=10.0.0.2:6789/0,c=10.0.0.5:6789/0,d=10.0.0.6:6789/0,e=10.0.0.3:6789/0}, > > election epoch 2584, quorum 0,1,2,3 a,b,c,e > > osdmap e82502: 50 osds: 48 up, 48 in" > > > > http://ceph.com/docs/master/rados/troubleshooting/troubleshooting-osd/#stopping-w-out-rebalancing > > > > If you have down OSDs that don't get marked out, that would certainly > > cause problems. Have you tried restarting the failed OSDs? > > > > What do the logs look like for osd.15 and osd.25? > > > > On Fri, May 17, 2013 at 1:31 AM, Olivier Bonvalet <ceph.list@xxxxxxxxx> wrote: > >> Hi, > >> > >> thanks for your answer. In fact I have several different problems, which > >> I tried to solve separatly : > >> > >> 1) I loose 2 OSD, and some pools have only 2 replicas. So some data was > >> lost. > >> 2) One monitor refuse the Cuttlefish upgrade, so I only have 4 of 5 > >> monitors running. > >> 3) I have 4 old inconsistent PG that I can't repair. > >> > >> > >> So the status : > >> > >> health HEALTH_ERR 15 pgs incomplete; 4 pgs inconsistent; 15 pgs stuck > >> inactive; 15 pgs stuck unclean; 1 near full osd(s); 19 scrub errors; > >> noout flag(s) set; 1 mons down, quorum 0,1,2,3 a,b,c,e > >> monmap e7: 5 mons at > >> {a=10.0.0.1:6789/0,b=10.0.0.2:6789/0,c=10.0.0.5:6789/0,d=10.0.0.6:6789/0,e=10.0.0.3:6789/0}, election epoch 2584, quorum 0,1,2,3 a,b,c,e > >> osdmap e82502: 50 osds: 48 up, 48 in > >> pgmap v12807617: 7824 pgs: 7803 active+clean, 1 active+clean > >> +scrubbing, 15 incomplete, 4 active+clean+inconsistent, 1 active+clean > >> +scrubbing+deep; 5676 GB data, 18948 GB used, 18315 GB / 37263 GB avail; > >> 137KB/s rd, 1852KB/s wr, 199op/s > >> mdsmap e1: 0/0/1 up > >> > >> > >> > >> The tree : > >> > >> # id weight type name up/down reweight > >> -8 14.26 root SSDroot > >> -27 8 datacenter SSDrbx2 > >> -26 8 room SSDs25 > >> -25 8 net SSD188-165-12 > >> -24 8 rack SSD25B09 > >> -23 8 host lyll > >> 46 2 osd.46 up 1 > >> 47 2 osd.47 up 1 > >> 48 2 osd.48 up 1 > >> 49 2 osd.49 up 1 > >> -10 4.26 datacenter SSDrbx3 > >> -12 2 room SSDs43 > >> -13 2 net SSD178-33-122 > >> -16 2 rack SSD43S01 > >> -17 2 host kaino > >> 42 1 osd.42 up 1 > >> 43 1 osd.43 up 1 > >> -22 2.26 room SSDs45 > >> -21 2.26 net SSD5-135-138 > >> -20 2.26 rack SSD45F01 > >> -19 2.26 host taman > >> 44 1.13 osd.44 up 1 > >> 45 1.13 osd.45 up 1 > >> -9 2 datacenter SSDrbx4 > >> -11 2 room SSDs52 > >> -14 2 net SSD176-31-226 > >> -15 2 rack SSD52B09 > >> -18 2 host dragan > >> 40 1 osd.40 up 1 > >> 41 1 osd.41 up 1 > >> -1 33.43 root SASroot > >> -100 15.9 datacenter SASrbx1 > >> -90 15.9 room SASs15 > >> -72 15.9 net SAS188-165-15 > >> -40 8 rack SAS15B01 > >> -3 8 host brontes > >> 0 1 osd.0 up 1 > >> 1 1 osd.1 up 1 > >> 2 1 osd.2 up 1 > >> 3 1 osd.3 up 1 > >> 4 1 osd.4 up 1 > >> 5 1 osd.5 up 1 > >> 6 1 osd.6 up 1 > >> 7 1 osd.7 up 1 > >> -41 7.9 rack SAS15B02 > >> -6 7.9 host alim > >> 24 1 osd.24 up 1 > >> 25 1 osd.25 down 0 > >> 26 1 osd.26 up 1 > >> 27 1 osd.27 up 1 > >> 28 1 osd.28 up 1 > >> 29 1 osd.29 up 1 > >> 30 1 osd.30 up 1 > >> 31 0.9 osd.31 up 1 > >> -101 17.53 datacenter SASrbx2 > >> -91 17.53 room SASs27 > >> -70 1.6 net SAS188-165-13 > >> -44 0 rack SAS27B04 > >> -7 0 host bul > >> -45 1.6 rack SAS27B06 > >> -4 1.6 host okko > >> 32 0.2 osd.32 up 1 > >> 33 0.2 osd.33 up 1 > >> 34 0.2 osd.34 up 1 > >> 35 0.2 osd.35 up 1 > >> 36 0.2 osd.36 up 1 > >> 37 0.2 osd.37 up 1 > >> 38 0.2 osd.38 up 1 > >> 39 0.2 osd.39 up 1 > >> -71 15.93 net SAS188-165-14 > >> -42 8 rack SAS27A03 > >> -5 8 host noburo > >> 8 1 osd.8 up 1 > >> 9 1 osd.9 up 1 > >> 18 1 osd.18 up 1 > >> 19 1 osd.19 up 1 > >> 20 1 osd.20 up 1 > >> 21 1 osd.21 up 1 > >> 22 1 osd.22 up 1 > >> 23 1 osd.23 up 1 > >> -43 7.93 rack SAS27A04 > >> -2 7.93 host keron > >> 10 0.97 osd.10 up 1 > >> 11 1 osd.11 up 1 > >> 12 1 osd.12 up 1 > >> 13 1 osd.13 up 1 > >> 14 0.98 osd.14 up 1 > >> 15 1 osd.15 down 0 > >> 16 0.98 osd.16 up 1 > >> 17 1 osd.17 up 1 > >> > >> > >> Here I have 2 roots : SSDroot and SASroot. All my OSD/PG problems are on > >> the SAS branch, and my CRUSH rules use per "net" replication. > >> > >> The osd.15 have a failling disk since long time, its data was correctly > >> moved (= OSD was out until the cluster obtain HEALTH_OK). > >> The osd.25 is a buggy OSD that I can't remove or change : if I balance > >> it's PG on other OSD, then this others OSD crash. That problem occur > >> before I loose the osd.19 : OSD was unable to mark that PG as > >> inconsistent since it was crashing during scrub. For me, all > >> inconsistencies come from this OSD. > >> The osd.19 was a failling disk, that I changed. > >> > >> > >> And the health detail : > >> > >> HEALTH_ERR 15 pgs incomplete; 4 pgs inconsistent; 15 pgs stuck inactive; > >> 15 pgs stuck unclean; 1 near full osd(s); 19 scrub errors; noout flag(s) > >> set; 1 mons down, quorum 0,1,2,3 a,b,c,e > >> pg 4.5c is stuck inactive since forever, current state incomplete, last > >> acting [19,30] > >> pg 8.71d is stuck inactive since forever, current state incomplete, last > >> acting [24,19] > >> pg 8.3fa is stuck inactive since forever, current state incomplete, last > >> acting [19,31] > >> pg 8.3e0 is stuck inactive since forever, current state incomplete, last > >> acting [31,19] > >> pg 8.56c is stuck inactive since forever, current state incomplete, last > >> acting [19,28] > >> pg 8.19f is stuck inactive since forever, current state incomplete, last > >> acting [31,19] > >> pg 8.792 is stuck inactive since forever, current state incomplete, last > >> acting [19,28] > >> pg 4.0 is stuck inactive since forever, current state incomplete, last > >> acting [28,19] > >> pg 8.78a is stuck inactive since forever, current state incomplete, last > >> acting [31,19] > >> pg 8.23e is stuck inactive since forever, current state incomplete, last > >> acting [32,13] > >> pg 8.2ff is stuck inactive since forever, current state incomplete, last > >> acting [6,19] > >> pg 8.5e2 is stuck inactive since forever, current state incomplete, last > >> acting [0,19] > >> pg 8.528 is stuck inactive since forever, current state incomplete, last > >> acting [31,19] > >> pg 8.20f is stuck inactive since forever, current state incomplete, last > >> acting [31,19] > >> pg 8.372 is stuck inactive since forever, current state incomplete, last > >> acting [19,24] > >> pg 4.5c is stuck unclean since forever, current state incomplete, last > >> acting [19,30] > >> pg 8.71d is stuck unclean since forever, current state incomplete, last > >> acting [24,19] > >> pg 8.3fa is stuck unclean since forever, current state incomplete, last > >> acting [19,31] > >> pg 8.3e0 is stuck unclean since forever, current state incomplete, last > >> acting [31,19] > >> pg 8.56c is stuck unclean since forever, current state incomplete, last > >> acting [19,28] > >> pg 8.19f is stuck unclean since forever, current state incomplete, last > >> acting [31,19] > >> pg 8.792 is stuck unclean since forever, current state incomplete, last > >> acting [19,28] > >> pg 4.0 is stuck unclean since forever, current state incomplete, last > >> acting [28,19] > >> pg 8.78a is stuck unclean since forever, current state incomplete, last > >> acting [31,19] > >> pg 8.23e is stuck unclean since forever, current state incomplete, last > >> acting [32,13] > >> pg 8.2ff is stuck unclean since forever, current state incomplete, last > >> acting [6,19] > >> pg 8.5e2 is stuck unclean since forever, current state incomplete, last > >> acting [0,19] > >> pg 8.528 is stuck unclean since forever, current state incomplete, last > >> acting [31,19] > >> pg 8.20f is stuck unclean since forever, current state incomplete, last > >> acting [31,19] > >> pg 8.372 is stuck unclean since forever, current state incomplete, last > >> acting [19,24] > >> pg 8.792 is incomplete, acting [19,28] > >> pg 8.78a is incomplete, acting [31,19] > >> pg 8.71d is incomplete, acting [24,19] > >> pg 8.5e2 is incomplete, acting [0,19] > >> pg 8.56c is incomplete, acting [19,28] > >> pg 8.528 is incomplete, acting [31,19] > >> pg 8.3fa is incomplete, acting [19,31] > >> pg 8.3e0 is incomplete, acting [31,19] > >> pg 8.372 is incomplete, acting [19,24] > >> pg 8.2ff is incomplete, acting [6,19] > >> pg 8.23e is incomplete, acting [32,13] > >> pg 8.20f is incomplete, acting [31,19] > >> pg 8.19f is incomplete, acting [31,19] > >> pg 3.7c is active+clean+inconsistent, acting [24,13,39] > >> pg 3.6b is active+clean+inconsistent, acting [28,23,5] > >> pg 4.5c is incomplete, acting [19,30] > >> pg 3.d is active+clean+inconsistent, acting [29,4,11] > >> pg 4.0 is incomplete, acting [28,19] > >> pg 3.1 is active+clean+inconsistent, acting [28,19,5] > >> osd.10 is near full at 85% > >> 19 scrub errors > >> noout flag(s) set > >> mon.d (rank 4) addr 10.0.0.6:6789/0 is down (out of quorum) > >> > >> > >> Pools 4 and 8 have only 2 replica, and pool 3 have 3 replica but > >> inconsistent data. > >> > >> Thanks in advance. > >> > >> Le vendredi 17 mai 2013 à 00:14 -0700, John Wilkins a écrit : > >>> If you can follow the documentation here: > >>> http://ceph.com/docs/master/rados/operations/monitoring-osd-pg/ and > >>> http://ceph.com/docs/master/rados/troubleshooting/ to provide some > >>> additional information, we may be better able to help you. > >>> > >>> For example, "ceph osd tree" would help us understand the status of > >>> your cluster a bit better. > >>> > >>> On Thu, May 16, 2013 at 10:32 PM, Olivier Bonvalet <ceph.list@xxxxxxxxx> wrote: > >>> > Le mercredi 15 mai 2013 à 00:15 +0200, Olivier Bonvalet a écrit : > >>> >> Hi, > >>> >> > >>> >> I have some PG in state down and/or incomplete on my cluster, because I > >>> >> loose 2 OSD and a pool was having only 2 replicas. So of course that > >>> >> data is lost. > >>> >> > >>> >> My problem now is that I can't retreive a "HEALTH_OK" status : if I try > >>> >> to remove, read or overwrite the corresponding RBD images, near all OSD > >>> >> hang (well... they don't do anything and requests stay in a growing > >>> >> queue, until the production will be done). > >>> >> > >>> >> So, what can I do to remove that corrupts images ? > >>> >> > >>> >> _______________________________________________ > >>> >> ceph-users mailing list > >>> >> ceph-users@xxxxxxxxxxxxxx > >>> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >>> >> > >>> > > >>> > Up. Nobody can help me on that problem ? > >>> > > >>> > Thanks, > >>> > > >>> > Olivier > >>> > > >>> > _______________________________________________ > >>> > ceph-users mailing list > >>> > ceph-users@xxxxxxxxxxxxxx > >>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >>> > >>> > >>> > >>> -- > >>> John Wilkins > >>> Senior Technical Writer > >>> Intank > >>> john.wilkins@xxxxxxxxxxx > >>> (415) 425-9599 > >>> http://inktank.com > >>> > >> > >> > > > > > > > > -- > > John Wilkins > > Senior Technical Writer > > Intank > > john.wilkins@xxxxxxxxxxx > > (415) 425-9599 > > http://inktank.com > > > > -- > John Wilkins > Senior Technical Writer > Intank > john.wilkins@xxxxxxxxxxx > (415) 425-9599 > http://inktank.com > -- > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html > -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html