Re: Ceph mon crash

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



2012/3/19 Greg Farnum <gregory.farnum@xxxxxxxxxxxxx>:
> On Monday, March 19, 2012 at 7:33 AM, ruslan usifov wrote:
>> Hello
>>
>> I have follow stack trace:
>>
>> #0 0xb77fa424 in __kernel_vsyscall ()
>> (gdb) bt
>> #0 0xb77fa424 in __kernel_vsyscall ()
>> #1 0xb77e98a0 in raise () from /lib/i386-linux-gnu/
>> libpthread.so.0
>> #2 0x08230f8b in ?? ()
>> #3 <signal handler called>
>> #4 0xb77fa424 in __kernel_vsyscall ()
>> #5 0xb70eae71 in raise () from /lib/i386-linux-gnu/libc.so.6
>> #6 0xb70ee34e in abort () from /lib/i386-linux-gnu/libc.so.6
>> #7 0xb73130b5 in __gnu_cxx::__verbose_terminate_handler() () from
>> /usr/lib/i386-linux-gnu/libstdc++.so.6
>> #8 0xb7310fa5 in ?? () from /usr/lib/i386-linux-gnu/libstdc++.so.6
>> #9 0xb7310fe2 in std::terminate() () from
>> /usr/lib/i386-linux-gnu/libstdc++.so.6
>> #10 0xb731114e in __cxa_throw () from /usr/lib/i386-linux-gnu/libstdc++.so.6
>> #11 0x0822f8c7 in ceph::__ceph_assert_fail(char const*, char const*,
>> int, char const*) ()
>> #12 0x081cf8a4 in MDSMap::get_health(std::basic_ostream<char,
>> std::char_traits<char> >&) const ()
>> #13 0x0811e8a7 in MDSMonitor::get_health(std::basic_ostream<char,
>> std::char_traits<char> >&) const ()
>> #14 0x080c4977 in Monitor::handle_command(MMonCommand*) ()
>> #15 0x080cf244 in Monitor::_ms_dispatch(Message*) ()
>> #16 0x080df1a4 in Monitor::ms_dispatch(Message*) ()
>> #17 0x081f706d in SimpleMessenger::dispatch_entry() ()
>> #18 0x080b27d2 in SimpleMessenger::DispatchThread::entry() ()
>> #19 0x081b5d81 in Thread::_entry_func(void*) ()
>> #20 0xb77e0e99 in start_thread () from /lib/i386-linux-gnu/libpthread.so.0
>> #21 0xb71919ee in clone () from /lib/i386-linux-gnu/libc.so.6
>
> Can you get the line number from frame 12? (f 12 <enter>, then just paste the output) Also the output of "ceph -s" if things are still running. The only assert I see in get_health() is that each "up" MDS be in mds_info, which really ought to be true….


Sorry but no, i use precompiled binaries from this
http://ceph.newdream.net/debian. Perhaps this helps, initialy i
configure all ceph services mon, mds, osd, but then i test only rdb
and remove all mds from cluster (3 vmware machines) throw follow
command:

ceph mds rm 1 (i write this lines by memory so can mistaken in syntax)

>
>> And when one mon crashes all other monitors in cluster will crashes
>> too:-((. So one time in cluster not any alive mons
>
> Yeah, this is because the crash is being triggered by a get_health command and it's trying it out on each monitor in turn as they fail.
> -Greg
>
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux