Re: Ceph Status - Segmentation Fault

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Jun 14, 2016 at 2:26 AM, Mathias Buresch
<mathias.buresch@xxxxxxxxxxxx> wrote:
> Hey,
>
> I opened an issue at tracker.ceph.com -> http://tracker.ceph.com/issues
> /16266

Hi Mathias,

Thanks!

I've added some information in that bug as I came across this same
issue working on something else and saw your bug this morning.

Cheers,
Brad

-----Original Message-----
> From: Brad Hubbard <bhubbard@xxxxxxxxxx>
> To: Mathias Buresch <mathias.buresch@xxxxxxxxxxxx>
> Cc: jspray@xxxxxxxxxx <jspray@xxxxxxxxxx>, ceph-users@xxxxxxxx <ceph-us
> ers@xxxxxxxx>
> Subject: Re:  Ceph Status - Segmentation Fault
> Date: Thu, 2 Jun 2016 09:50:20 +1000
>
> Could this be the call in RotatingKeyRing::get_secret() failing?
>
> Mathias, I'd suggest opening a tracker for this with the information in
> your last post and let us know the number here.
> Cheers,
> Brad
>
> On Wed, Jun 1, 2016 at 3:15 PM, Mathias Buresch <mathias.buresch@xxxxxx
> ra.net> wrote:
>> Hi,
>>
>> here is the output including --debug-auth=20. Does this help?
>>
>> (gdb) run /usr/bin/ceph status --debug-monc=20 --debug-ms=20 --debug-
>> rados=20 --debug-auth=20
>> Starting program: /usr/bin/python /usr/bin/ceph status --debug-
>> monc=20
>> --debug-ms=20 --debug-rados=20 --debug-auth=20
>> [Thread debugging using libthread_db enabled]
>> Using host libthread_db library "/lib/x86_64-linux-
>> gnu/libthread_db.so.1".
>> [New Thread 0x7ffff10f5700 (LWP 2210)]
>> [New Thread 0x7ffff08f4700 (LWP 2211)]
>> [Thread 0x7ffff10f5700 (LWP 2210) exited]
>> [New Thread 0x7ffff10f5700 (LWP 2212)]
>> [Thread 0x7ffff10f5700 (LWP 2212) exited]
>> [New Thread 0x7ffff10f5700 (LWP 2213)]
>> [Thread 0x7ffff10f5700 (LWP 2213) exited]
>> [New Thread 0x7ffff10f5700 (LWP 2233)]
>> [Thread 0x7ffff10f5700 (LWP 2233) exited]
>> [New Thread 0x7ffff10f5700 (LWP 2236)]
>> [Thread 0x7ffff10f5700 (LWP 2236) exited]
>> [New Thread 0x7ffff10f5700 (LWP 2237)]
>> [Thread 0x7ffff10f5700 (LWP 2237) exited]
>> [New Thread 0x7ffff10f5700 (LWP 2238)]
>> [New Thread 0x7fffeb885700 (LWP 2240)]
>> 2016-06-01 07:12:55.656336 7ffff10f5700 10 monclient(hunting):
>> build_initial_monmap
>> 2016-06-01 07:12:55.656440 7ffff10f5700  1 librados: starting msgr at
>> :/0
>> 2016-06-01 07:12:55.656446 7ffff10f5700  1 librados: starting
>> objecter
>> [New Thread 0x7fffeb084700 (LWP 2241)]
>> 2016-06-01 07:12:55.657552 7ffff10f5700 10 -- :/0 ready :/0
>> [New Thread 0x7fffea883700 (LWP 2242)]
>> [New Thread 0x7fffea082700 (LWP 2245)]
>> 2016-06-01 07:12:55.659548 7ffff10f5700  1 -- :/0 messenger.start
>> [New Thread 0x7fffe9881700 (LWP 2248)]
>> 2016-06-01 07:12:55.660530 7ffff10f5700  1 librados: setting wanted
>> keys
>> 2016-06-01 07:12:55.660539 7ffff10f5700  1 librados: calling
>> monclient
>> init
>> 2016-06-01 07:12:55.660540 7ffff10f5700 10 monclient(hunting): init
>> 2016-06-01 07:12:55.660550 7ffff10f5700  5 adding auth protocol:
>> cephx
>> 2016-06-01 07:12:55.660552 7ffff10f5700 10 monclient(hunting):
>> auth_supported 2 method cephx
>> 2016-06-01 07:12:55.660532 7fffe9881700 10 -- :/1337675866
>> reaper_entry
>> start
>> 2016-06-01 07:12:55.660570 7fffe9881700 10 -- :/1337675866 reaper
>> 2016-06-01 07:12:55.660572 7fffe9881700 10 -- :/1337675866 reaper
>> done
>> 2016-06-01 07:12:55.660733 7ffff10f5700  2 auth: KeyRing::load:
>> loaded
>> key file /etc/ceph/ceph.client.admin.keyring
>> [New Thread 0x7fffe9080700 (LWP 2251)]
>> [New Thread 0x7fffe887f700 (LWP 2252)]
>> 2016-06-01 07:12:55.662754 7ffff10f5700 10 monclient(hunting):
>> _reopen_session rank -1 name
>> 2016-06-01 07:12:55.662764 7ffff10f5700 10 -- :/1337675866
>> connect_rank
>> to 62.176.141.181:6789/0, creating pipe and registering
>> [New Thread 0x7fffe3fff700 (LWP 2255)]
>> 2016-06-01 07:12:55.663789 7ffff10f5700 10 -- :/1337675866 >>
>> 62.176.141.181:6789/0 pipe(0x7fffec064010 sd=-1 :0 s=1 pgs=0 cs=0 l=1
>> c=0x7fffec05aa30).register_pipe
>> 2016-06-01 07:12:55.663819 7ffff10f5700 10 -- :/1337675866
>> get_connection mon.0 62.176.141.181:6789/0 new 0x7fffec064010
>> 2016-06-01 07:12:55.663790 7fffe3fff700 10 -- :/1337675866 >>
>> 62.176.141.181:6789/0 pipe(0x7fffec064010 sd=-1 :0 s=1 pgs=0 cs=0 l=1
>> c=0x7fffec05aa30).writer: state = connecting policy.server=0
>> 2016-06-01 07:12:55.663830 7fffe3fff700 10 -- :/1337675866 >>
>> 62.176.141.181:6789/0 pipe(0x7fffec064010 sd=-1 :0 s=1 pgs=0 cs=0 l=1
>> c=0x7fffec05aa30).connect 0
>> 2016-06-01 07:12:55.663841 7ffff10f5700 10 monclient(hunting): picked
>> mon.pix01 con 0x7fffec05aa30 addr 62.176.141.181:6789/0
>> 2016-06-01 07:12:55.663847 7ffff10f5700 20 -- :/1337675866
>> send_keepalive con 0x7fffec05aa30, have pipe.
>> 2016-06-01 07:12:55.663850 7fffe3fff700 10 -- :/1337675866 >>
>> 62.176.141.181:6789/0 pipe(0x7fffec064010 sd=3 :0 s=1 pgs=0 cs=0 l=1
>> c=0x7fffec05aa30).connecting to 62.176.141.181:6789/0
>> 2016-06-01 07:12:55.663863 7ffff10f5700 10 monclient(hunting):
>> _send_mon_message to mon.pix01 at 62.176.141.181:6789/0
>> 2016-06-01 07:12:55.663866 7ffff10f5700  1 -- :/1337675866 -->
>> 62.176.141.181:6789/0 -- auth(proto 0 30 bytes epoch 0) v1 -- ?+0
>> 0x7fffec060450 con 0x7fffec05aa30
>> 2016-06-01 07:12:55.663870 7ffff10f5700 20 -- :/1337675866
>> submit_message auth(proto 0 30 bytes epoch 0) v1 remote,
>> 62.176.141.181:6789/0, have pipe.
>> 2016-06-01 07:12:55.663874 7ffff10f5700 10 monclient(hunting):
>> renew_subs
>> 2016-06-01 07:12:55.663877 7ffff10f5700 10 monclient(hunting):
>> authenticate will time out at 2016-06-01 07:17:55.663876
>> 2016-06-01 07:12:55.664115 7fffe3fff700 20 -- :/1337675866 >>
>> 62.176.141.181:6789/0 pipe(0x7fffec064010 sd=3 :41128 s=1 pgs=0 cs=0
>> l=1 c=0x7fffec05aa30).connect read peer addr 62.176.141.181:6789/0 on
>> socket 3
>> 2016-06-01 07:12:55.664135 7fffe3fff700 20 -- :/1337675866 >>
>> 62.176.141.181:6789/0 pipe(0x7fffec064010 sd=3 :41128 s=1 pgs=0 cs=0
>> l=1 c=0x7fffec05aa30).connect peer addr for me is
>> 62.176.141.181:41128/0
>> 2016-06-01 07:12:55.664143 7fffe3fff700  1 --
>> 62.176.141.181:0/1337675866 learned my addr
>> 62.176.141.181:0/1337675866
>> 2016-06-01 07:12:55.664177 7fffe3fff700 10 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=1 pgs=0 cs=0 l=1
>> c=0x7fffec05aa30).connect sent my addr 62.176.141.181:0/1337675866
>> 2016-06-01 07:12:55.664193 7fffe3fff700 10 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=1 pgs=0 cs=0 l=1
>> c=0x7fffec05aa30).connect sending gseq=1 cseq=0 proto=15
>> 2016-06-01 07:12:55.664207 7fffe3fff700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=1 pgs=0 cs=0 l=1
>> c=0x7fffec05aa30).connect wrote (self +) cseq, waiting for reply
>> 2016-06-01 07:12:55.664268 7fffe3fff700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=1 pgs=0 cs=0 l=1
>> c=0x7fffec05aa30).connect got reply tag 1 connect_seq 1 global_seq
>> 339278 proto 15 flags 1 features 55169095435288575
>> 2016-06-01 07:12:55.664285 7fffe3fff700 10 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).connect success 1, lossy = 1, features
>> 55169095435288575
>> 2016-06-01 07:12:55.664301 7fffe3fff700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).connect starting reader
>> [New Thread 0x7fffe3efe700 (LWP 2258)]
>> 2016-06-01 07:12:55.665288 7fffe3fff700 10 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).writer: state = open policy.server=0
>> 2016-06-01 07:12:55.665307 7fffe3fff700 10 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).write_keepalive2 14 2016-06-01 07:12:55.665306
>> 2016-06-01 07:12:55.665310 7fffe3efe700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).reader reading tag...
>> 2016-06-01 07:12:55.665347 7fffe3fff700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).writer encoding 1 features 55169095435288575
>> 0x7fffec060450 auth(proto 0 30 bytes epoch 0) v1
>> 2016-06-01 07:12:55.665376 7fffe3fff700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).writer no session security
>> 2016-06-01 07:12:55.665383 7fffe3fff700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).writer sending 1 0x7fffec060450
>> 2016-06-01 07:12:55.665384 7fffe3efe700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).reader got KEEPALIVE_ACK
>> 2016-06-01 07:12:55.665394 7fffe3efe700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).reader reading tag...
>> 2016-06-01 07:12:55.665401 7fffe3fff700 10 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).writer: state = open policy.server=0
>> 2016-06-01 07:12:55.665410 7fffe3fff700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).writer sleeping
>> 2016-06-01 07:12:55.665706 7fffe3efe700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).reader got ACK
>> 2016-06-01 07:12:55.665746 7fffe3efe700 15 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).reader got ack seq 1
>> 2016-06-01 07:12:55.665752 7fffe3efe700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).reader reading tag...
>> 2016-06-01 07:12:55.665764 7fffe3efe700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).reader got MSG
>> 2016-06-01 07:12:55.665769 7fffe3efe700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).reader got envelope type=4 src mon.0 front=340
>> data=0
>> off 0
>> 2016-06-01 07:12:55.665787 7fffe3efe700 10 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).reader wants 340 from dispatch throttler
>> 0/104857600
>> 2016-06-01 07:12:55.665797 7fffe3efe700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).reader got front 340
>> 2016-06-01 07:12:55.665801 7fffe3efe700 10 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).aborted = 0
>> 2016-06-01 07:12:55.665804 7fffe3efe700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).reader got 340 + 0 + 0 byte message
>> 2016-06-01 07:12:55.665825 7fffe3efe700 10 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).No session security set
>> 2016-06-01 07:12:55.665834 7fffe3efe700 10 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).reader got message 1 0x7fffd0001cb0 mon_map magic:
>> 0
>> v1
>> 2016-06-01 07:12:55.665845 7fffe3efe700 20 --
>> 62.176.141.181:0/1337675866 queue 0x7fffd0001cb0 prio 196
>> 2016-06-01 07:12:55.665853 7fffe3efe700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).reader reading tag...
>> 2016-06-01 07:12:55.665862 7fffe3fff700 10 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).writer: state = open policy.server=0
>> 2016-06-01 07:12:55.665874 7fffe3efe700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).reader got MSG
>> 2016-06-01 07:12:55.665879 7fffe3efe700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).reader got envelope type=18 src mon.0 front=33
>> data=0
>> off 0
>> 2016-06-01 07:12:55.665879 7fffe3fff700 10 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).write_ack 1
>> 2016-06-01 07:12:55.665883 7fffe3efe700 10 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).reader wants 33 from dispatch throttler
>> 340/104857600
>> 2016-06-01 07:12:55.665888 7fffe3efe700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).reader got front 33
>> 2016-06-01 07:12:55.665891 7fffe3efe700 10 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).aborted = 0
>> 2016-06-01 07:12:55.665891 7fffe3fff700 10 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).writer: state = open policy.server=0
>> 2016-06-01 07:12:55.665895 7fffe3efe700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).reader got 33 + 0 + 0 byte message
>> 2016-06-01 07:12:55.665896 7fffe3fff700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).writer sleeping
>> 2016-06-01 07:12:55.665916 7fffe3efe700 10 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).No session security set
>> 2016-06-01 07:12:55.665922 7fffe3efe700 10 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).reader got message 2 0x7fffd0002f20
>> auth_reply(proto
>> 2 0 (0) Success) v1
>> 2016-06-01 07:12:55.665944 7fffe3efe700 20 --
>> 62.176.141.181:0/1337675866 queue 0x7fffd0002f20 prio 196
>> 2016-06-01 07:12:55.665950 7fffe3efe700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).reader reading tag...
>> 2016-06-01 07:12:55.665891 7fffea883700  1 --
>> 62.176.141.181:0/1337675866 <== mon.0 62.176.141.181:6789/0 1 ====
>> mon_map magic: 0 v1 ==== 340+0+0 (3213884171 0 0) 0x7fffd0001cb0 con
>> 0x7fffec05aa30
>> 2016-06-01 07:12:55.665953 7fffe3fff700 10 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).writer: state = open policy.server=0
>> 2016-06-01 07:12:55.665960 7fffea883700 10 monclient(hunting):
>> handle_monmap mon_map magic: 0 v1
>> 2016-06-01 07:12:55.665960 7fffe3fff700 10 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).write_ack 2
>> 2016-06-01 07:12:55.665966 7fffe3fff700 10 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).writer: state = open policy.server=0
>> 2016-06-01 07:12:55.665971 7fffea883700 10 monclient(hunting):  got
>> monmap 1, mon.pix01 is now rank 0
>> 2016-06-01 07:12:55.665970 7fffe3fff700 20 --
>> 62.176.141.181:0/1337675866 >> 62.176.141.181:6789/0
>> pipe(0x7fffec064010 sd=3 :41128 s=2 pgs=339278 cs=1 l=1
>> c=0x7fffec05aa30).writer sleeping
>> 2016-06-01 07:12:55.665972 7fffea883700 10 monclient(hunting): dump:
>> epoch 1
>> fsid 28af67eb-4060-4770-ac1d-d2be493877af
>> last_changed 2014-11-12 15:44:27.182395
>> created 2014-11-12 15:44:27.182395
>> 0: 62.176.141.181:6789/0 mon.pix01
>> 1: 62.176.141.182:6789/0 mon.pix02
>>
>> 2016-06-01 07:12:55.665988 7fffea883700 10 --
>> 62.176.141.181:0/1337675866 dispatch_throttle_release 340 to dispatch
>> throttler 373/104857600
>> 2016-06-01 07:12:55.665992 7fffea883700 20 --
>> 62.176.141.181:0/1337675866 done calling dispatch on 0x7fffd0001cb0
>> 2016-06-01 07:12:55.665997 7fffea883700  1 --
>> 62.176.141.181:0/1337675866 <== mon.0 62.176.141.181:6789/0 2 ====
>> auth_reply(proto 2 0 (0) Success) v1 ==== 33+0+0 (3918039325 0 0)
>> 0x7fffd0002f20 con 0x7fffec05aa30
>> 2016-06-01 07:12:55.666015 7fffea883700 10 cephx: set_have_need_key
>> no
>> handler for service mon
>> 2016-06-01 07:12:55.666016 7fffea883700 10 cephx: set_have_need_key
>> no
>> handler for service osd
>> 2016-06-01 07:12:55.666017 7fffea883700 10 cephx: set_have_need_key
>> no
>> handler for service auth
>> 2016-06-01 07:12:55.666018 7fffea883700 10 cephx: validate_tickets
>> want
>> 37 have 0 need 37
>> 2016-06-01 07:12:55.666020 7fffea883700 10 monclient(hunting): my
>> global_id is 3511432
>> 2016-06-01 07:12:55.666022 7fffea883700 10 cephx client:
>> handle_response ret = 0
>> 2016-06-01 07:12:55.666023 7fffea883700 10 cephx client:  got initial
>> server challenge 3112857369079243605
>> 2016-06-01 07:12:55.666025 7fffea883700 10 cephx client:
>> validate_tickets: want=37 need=37 have=0
>> 2016-06-01 07:12:55.666026 7fffea883700 10 cephx: set_have_need_key
>> no
>> handler for service mon
>> 2016-06-01 07:12:55.666027 7fffea883700 10 cephx: set_have_need_key
>> no
>> handler for service osd
>> 2016-06-01 07:12:55.666030 7fffea883700 10 cephx: set_have_need_key
>> no
>> handler for service auth
>> 2016-06-01 07:12:55.666030 7fffea883700 10 cephx: validate_tickets
>> want
>> 37 have 0 need 37
>> 2016-06-01 07:12:55.666031 7fffea883700 10 cephx client: want=37
>> need=37 have=0
>> 2016-06-01 07:12:55.666034 7fffea883700 10 cephx client:
>> build_request
>>
>> Program received signal SIGSEGV, Segmentation fault.
>> [Switching to Thread 0x7fffea883700 (LWP 2242)]
>> 0x00007ffff3141a57 in encrypt (cct=<optimized out>,
>> error=0x7fffea882280, out=..., in=..., this=0x7fffea882470)
>>     at auth/cephx/../Crypto.h:110
>> 110     auth/cephx/../Crypto.h: No such file or directory.
>> (gdb) bt
>> #0  0x00007ffff3141a57 in encrypt (cct=<optimized out>,
>> error=0x7fffea882280, out=..., in=..., this=0x7fffea882470)
>>     at auth/cephx/../Crypto.h:110
>> #1  encode_encrypt_enc_bl<CephXChallengeBlob> (cct=<optimized out>,
>> error="", out=..., key=..., t=<synthetic pointer>)
>>     at auth/cephx/CephxProtocol.h:464
>> #2  encode_encrypt<CephXChallengeBlob> (cct=<optimized out>,
>> error="",
>> out=..., key=..., t=<synthetic pointer>)
>>     at auth/cephx/CephxProtocol.h:489
>> #3  cephx_calc_client_server_challenge (cct=<optimized out>,
>> secret=..., server_challenge=3112857369079243605,
>>     client_challenge=12899511428024786235, key=key@entry=0x7fffea8824
>> a8
>> , ret="") at auth/cephx/CephxProtocol.cc:36
>> #4  0x00007ffff313aff4 in CephxClientHandler::build_request
>> (this=0x7fffd4001520, bl=...) at auth/cephx/CephxClientHandler.cc:53
>> #5  0x00007ffff2fe4a79 in MonClient::handle_auth (this=this@entry=0x7
>> ff
>> fec006b70, m=m@entry=0x7fffd0002f20) at mon/MonClient.cc:510
>> #6  0x00007ffff2fe6507 in MonClient::ms_dispatch
>> (this=0x7fffec006b70,
>> m=0x7fffd0002f20) at mon/MonClient.cc:277
>> #7  0x00007ffff30d5dc9 in ms_deliver_dispatch (m=0x7fffd0002f20,
>> this=0x7fffec055410) at ./msg/Messenger.h:582
>> #8  DispatchQueue::entry (this=0x7fffec0555d8) at
>> msg/simple/DispatchQueue.cc:185
>> #9  0x00007ffff31023bd in DispatchQueue::DispatchThread::entry
>> (this=<optimized out>) at msg/simple/DispatchQueue.h:103
>> #10 0x00007ffff7bc4182 in start_thread () from /lib/x86_64-linux-
>> gnu/libpthread.so.0
>> #11 0x00007ffff78f147d in clone () from /lib/x86_64-linux-
>> gnu/libc.so.6
>>
>>
>> Best regards
>> Mathias-----Original Message-----
>> From: Brad Hubbard <bhubbard@xxxxxxxxxx>
>> To: jspray@xxxxxxxxxx
>> Cc: ceph-users@xxxxxxxx, Mathias Buresch <mathias.buresch@xxxxxxxxxxx
>> t>
>> Subject: Re:  Ceph Status - Segmentation Fault
>> Date: Wed, 25 May 2016 19:22:03 -0400
>>
>> Hi John,
>>
>> This looks a lot like http://tracker.ceph.com/issues/12417 which is,
>> of
>> course, fixed.
>>
>> Worth gathering debug-auth=20 ? Maybe on the MON end as well?
>>
>> Cheers,
>> Brad
>>
>>
>> ----- Original Message -----
>> >
>> > From: "Mathias Buresch" <mathias.buresch@xxxxxxxxxxxx>
>> > To: jspray@xxxxxxxxxx
>> > Cc: ceph-users@xxxxxxxx
>> > Sent: Thursday, 26 May, 2016 12:57:47 AM
>> > Subject: Re:  Ceph Status - Segmentation Fault
>> >
>> > There wasnt a package ceph-debuginfo available (Maybe bc I am
>> running
>> > Ubuntu). Have installed those:
>> >
>> >  * ceph-dbg
>> >  * librados2-dbg
>> >
>> > There would be also ceph-mds-dbg and ceph-fs-common-dbg and so..
>> >
>> > But now there are more information provided by the gdb output :)
>> >
>> > (gdb) run /usr/bin/ceph status --debug-monc=20 --debug-ms=20 --
>> debug-
>> > rados=20
>> > Starting program: /usr/bin/python /usr/bin/ceph status --debug-
>> > monc=20
>> > --debug-ms=20 --debug-rados=20
>> > [Thread debugging using libthread_db enabled]
>> > Using host libthread_db library "/lib/x86_64-linux-
>> > gnu/libthread_db.so.1".
>> > [New Thread 0x7ffff10f5700 (LWP 26739)]
>> > [New Thread 0x7ffff08f4700 (LWP 26740)]
>> > [Thread 0x7ffff10f5700 (LWP 26739) exited]
>> > [New Thread 0x7ffff10f5700 (LWP 26741)]
>> > [Thread 0x7ffff10f5700 (LWP 26741) exited]
>> > [New Thread 0x7ffff10f5700 (LWP 26742)]
>> > [Thread 0x7ffff10f5700 (LWP 26742) exited]
>> > [New Thread 0x7ffff10f5700 (LWP 26743)]
>> > [Thread 0x7ffff10f5700 (LWP 26743) exited]
>> > [New Thread 0x7ffff10f5700 (LWP 26744)]
>> > [Thread 0x7ffff10f5700 (LWP 26744) exited]
>> > [New Thread 0x7ffff10f5700 (LWP 26745)]
>> > [Thread 0x7ffff10f5700 (LWP 26745) exited]
>> > [New Thread 0x7ffff10f5700 (LWP 26746)]
>> > [New Thread 0x7fffeb885700 (LWP 26747)]
>> > 2016-05-25 16:55:30.929131 7ffff10f5700 10 monclient(hunting):
>> > build_initial_monmap
>> > 2016-05-25 16:55:30.929221 7ffff10f5700  1 librados: starting msgr
>> at
>> > :/0
>> > 2016-05-25 16:55:30.929226 7ffff10f5700  1 librados: starting
>> > objecter
>> > [New Thread 0x7fffeb084700 (LWP 26748)]
>> > 2016-05-25 16:55:30.930288 7ffff10f5700 10 -- :/0 ready :/0
>> > [New Thread 0x7fffea883700 (LWP 26749)]
>> > [New Thread 0x7fffea082700 (LWP 26750)]
>> > 2016-05-25 16:55:30.932251 7ffff10f5700  1 -- :/0 messenger.start
>> > [New Thread 0x7fffe9881700 (LWP 26751)]
>> > 2016-05-25 16:55:30.933277 7ffff10f5700  1 librados: setting wanted
>> > keys
>> > 2016-05-25 16:55:30.933287 7ffff10f5700  1 librados: calling
>> > monclient
>> > init
>> > 2016-05-25 16:55:30.933289 7ffff10f5700 10 monclient(hunting): init
>> > 2016-05-25 16:55:30.933279 7fffe9881700 10 -- :/3663984981
>> > reaper_entry
>> > start
>> > 2016-05-25 16:55:30.933300 7ffff10f5700 10 monclient(hunting):
>> > auth_supported 2 method cephx
>> > 2016-05-25 16:55:30.933303 7fffe9881700 10 -- :/3663984981 reaper
>> > 2016-05-25 16:55:30.933305 7fffe9881700 10 -- :/3663984981 reaper
>> > done
>> > [New Thread 0x7fffe9080700 (LWP 26752)]
>> > [New Thread 0x7fffe887f700 (LWP 26753)]
>> > 2016-05-25 16:55:30.935485 7ffff10f5700 10 monclient(hunting):
>> > _reopen_session rank -1 name
>> > 2016-05-25 16:55:30.935495 7ffff10f5700 10 -- :/3663984981
>> > connect_rank
>> > to 62.176.141.181:6789/0, creating pipe and registering
>> > [New Thread 0x7fffe3fff700 (LWP 26754)]
>> > 2016-05-25 16:55:30.936556 7ffff10f5700 10 -- :/3663984981 >>
>> > 62.176.141.181:6789/0 pipe(0x7fffec064010 sd=-1 :0 s=1 pgs=0 cs=0
>> l=1
>> > c=0x7fffec05aa30).register_pipe
>> > 2016-05-25 16:55:30.936573 7ffff10f5700 10 -- :/3663984981
>> > get_connection mon.0 62.176.141.181:6789/0 new 0x7fffec064010
>> > 2016-05-25 16:55:30.936557 7fffe3fff700 10 -- :/3663984981 >>
>> > 62.176.141.181:6789/0 pipe(0x7fffec064010 sd=-1 :0 s=1 pgs=0 cs=0
>> l=1
>> > c=0x7fffec05aa30).writer: state = connecting policy.server=0
>> > 2016-05-25 16:55:30.936583 7fffe3fff700 10 -- :/3663984981 >>
>> > 62.176.141.181:6789/0 pipe(0x7fffec064010 sd=-1 :0 s=1 pgs=0 cs=0
>> l=1
>> > c=0x7fffec05aa30).connect 0
>> > 2016-05-25 16:55:30.936594 7ffff10f5700 10 monclient(hunting):
>> picked
>> > mon.pix01 con 0x7fffec05aa30 addr 62.176.141.181:6789/0
>> > 2016-05-25 16:55:30.936600 7ffff10f5700 20 -- :/3663984981
>> > send_keepalive con 0x7fffec05aa30, have pipe.
>> > 2016-05-25 16:55:30.936603 7fffe3fff700 10 -- :/3663984981 >>
>> > 62.176.141.181:6789/0 pipe(0x7fffec064010 sd=3 :0 s=1 pgs=0 cs=0
>> l=1
>> > c=0x7fffec05aa30).connecting to 62.176.141.181:6789/0
>> > 2016-05-25 16:55:30.936615 7ffff10f5700 10 monclient(hunting):
>> > _send_mon_message to mon.pix01 at 62.176.141.181:6789/0
>> > 2016-05-25 16:55:30.936618 7ffff10f5700  1 -- :/3663984981 -->
>> > 62.176.141.181:6789/0 -- auth(proto 0 30 bytes epoch 0) v1 -- ?+0
>> > 0x7fffec060450 con 0x7fffec05aa30
>> > 2016-05-25 16:55:30.936623 7ffff10f5700 20 -- :/3663984981
>> > submit_message auth(proto 0 30 bytes epoch 0) v1 remote,
>> > 62.176.141.181:6789/0, have pipe.
>> > 2016-05-25 16:55:30.936626 7ffff10f5700 10 monclient(hunting):
>> > renew_subs
>> > 2016-05-25 16:55:30.936630 7ffff10f5700 10 monclient(hunting):
>> > authenticate will time out at 2016-05-25 17:00:30.936629
>> > 2016-05-25 16:55:30.936867 7fffe3fff700 20 -- :/3663984981 >>
>> > 62.176.141.181:6789/0 pipe(0x7fffec064010 sd=3 :38763 s=1 pgs=0
>> cs=0
>> > l=1 c=0x7fffec05aa30).connect read peer addr 62.176.141.181:6789/0
>> on
>> > socket 3
>> > 2016-05-25 16:55:30.936882 7fffe3fff700 20 -- :/3663984981 >>
>> > 62.176.141.181:6789/0 pipe(0x7fffec064010 sd=3 :38763 s=1 pgs=0
>> cs=0
>> > l=1 c=0x7fffec05aa30).connect peer addr for me is
>> > 62.176.141.181:38763/0
>> > 2016-05-25 16:55:30.936887 7fffe3fff700  1 --
>> > 62.176.141.181:0/3663984981 learned my addr
>> > 62.176.141.181:0/3663984981
>> > 2016-05-25 16:55:30.936914 7fffe3fff700 10 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=1 pgs=0 cs=0 l=1
>> > c=0x7fffec05aa30).connect sent my addr 62.176.141.181:0/3663984981
>> > 2016-05-25 16:55:30.936922 7fffe3fff700 10 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=1 pgs=0 cs=0 l=1
>> > c=0x7fffec05aa30).connect sending gseq=1 cseq=0 proto=15
>> > 2016-05-25 16:55:30.936932 7fffe3fff700 20 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=1 pgs=0 cs=0 l=1
>> > c=0x7fffec05aa30).connect wrote (self +) cseq, waiting for reply
>> > 2016-05-25 16:55:30.937015 7fffe3fff700 20 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=1 pgs=0 cs=0 l=1
>> > c=0x7fffec05aa30).connect got reply tag 1 connect_seq 1 global_seq
>> > 327867 proto 15 flags 1 features 55169095435288575
>> > 2016-05-25 16:55:30.937035 7fffe3fff700 10 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).connect success 1, lossy = 1, features
>> > 55169095435288575
>> > 2016-05-25 16:55:30.937046 7fffe3fff700 20 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).connect starting reader
>> > [New Thread 0x7fffe3efe700 (LWP 26757)]
>> > 2016-05-25 16:55:30.938118 7fffe3fff700 10 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).writer: state = open policy.server=0
>> > 2016-05-25 16:55:30.938127 7fffe3fff700 10 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).write_keepalive2 14 2016-05-25 16:55:30.938126
>> > 2016-05-25 16:55:30.938161 7fffe3fff700 20 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).writer encoding 1 features 55169095435288575
>> > 0x7fffec060450 auth(proto 0 30 bytes epoch 0) v1
>> > 2016-05-25 16:55:30.938185 7fffe3fff700 20 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).writer no session security
>> > 2016-05-25 16:55:30.938197 7fffe3fff700 20 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).writer sending 1 0x7fffec060450
>> > 2016-05-25 16:55:30.938216 7fffe3fff700 10 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).writer: state = open policy.server=0
>> > 2016-05-25 16:55:30.938222 7fffe3fff700 20 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).writer sleeping
>> > 2016-05-25 16:55:30.938201 7fffe3efe700 20 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).reader reading tag...
>> > 2016-05-25 16:55:30.938238 7fffe3efe700 20 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).reader got KEEPALIVE_ACK
>> > 2016-05-25 16:55:30.938243 7fffe3efe700 20 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).reader reading tag...
>> > 2016-05-25 16:55:30.938368 7fffe3efe700 20 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).reader got ACK
>> > 2016-05-25 16:55:30.938378 7fffe3efe700 15 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).reader got ack seq 1
>> > 2016-05-25 16:55:30.938382 7fffe3efe700 20 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).reader reading tag...
>> > 2016-05-25 16:55:30.938388 7fffe3efe700 20 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).reader got MSG
>> > 2016-05-25 16:55:30.938392 7fffe3efe700 20 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).reader got envelope type=4 src mon.0 front=340
>> > data=0
>> > off 0
>> > 2016-05-25 16:55:30.938399 7fffe3efe700 10 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).reader wants 340 from dispatch throttler
>> > 0/104857600
>> > 2016-05-25 16:55:30.938406 7fffe3efe700 20 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).reader got front 340
>> > 2016-05-25 16:55:30.938409 7fffe3efe700 10 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).aborted = 0
>> > 2016-05-25 16:55:30.938413 7fffe3efe700 20 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).reader got 340 + 0 + 0 byte message
>> > 2016-05-25 16:55:30.938427 7fffe3efe700 10 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).No session security set
>> > 2016-05-25 16:55:30.938434 7fffe3efe700 10 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).reader got message 1 0x7fffd0001cb0 mon_map
>> magic:
>> > 0
>> > v1
>> > 2016-05-25 16:55:30.938442 7fffe3efe700 20 --
>> > 62.176.141.181:0/3663984981 queue 0x7fffd0001cb0 prio 196
>> > 2016-05-25 16:55:30.938450 7fffe3efe700 20 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).reader reading tag...
>> > 2016-05-25 16:55:30.938453 7fffe3fff700 10 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).writer: state = open policy.server=0
>> > 2016-05-25 16:55:30.938464 7fffe3fff700 10 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).write_ack 1
>> > 2016-05-25 16:55:30.938467 7fffe3efe700 20 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).reader got MSG
>> > 2016-05-25 16:55:30.938471 7fffe3fff700 10 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).writer: state = open policy.server=0
>> > 2016-05-25 16:55:30.938472 7fffe3efe700 20 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).reader got envelope type=18 src mon.0 front=33
>> > data=0
>> > off 0
>> > 2016-05-25 16:55:30.938475 7fffe3fff700 20 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).writer sleeping
>> > 2016-05-25 16:55:30.938476 7fffe3efe700 10 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).reader wants 33 from dispatch throttler
>> > 340/104857600
>> > 2016-05-25 16:55:30.938456 7fffea883700  1 --
>> > 62.176.141.181:0/3663984981 <== mon.0 62.176.141.181:6789/0 1 ====
>> > mon_map magic: 0 v1 ==== 340+0+0 (3213884171 0 0) 0x7fffd0001cb0
>> con
>> > 0x7fffec05aa30
>> > 2016-05-25 16:55:30.938481 7fffe3efe700 20 --
>> > 62.176.141.181:0/3663984981 >> 62.176.141.181:6789/0
>> > pipe(0x7fffec064010 sd=3 :38763 s=2 pgs=327867 cs=1 l=1
>> > c=0x7fffec05aa30).reader got front 33
>> > 2016-05-25 16:55:30.938484 7fffea883700 10 monclient(hunting):
>> > handle_monmap mon_map magic: 0 v1
>> > 2016-05-25 16:55:30.938485 7fffe3efe700 10
>> > Program received signal SIGSEGV, Segmentation fault.
>> > [Switching to Thread 0x7fffea883700 (LWP 26749)]
>> > 0x00007ffff3141a57 in encrypt (cct=<optimized out>,
>> > error=0x7fffea882280, out=..., in=..., this=0x7fffea882470) at
>> > auth/cephx/../Crypto.h:110
>> > 110   auth/cephx/../Crypto.h: No such file or directory.
>> > (gdb) bt
>> > #0  0x00007ffff3141a57 in encrypt (cct=<optimized out>,
>> > error=0x7fffea882280, out=..., in=..., this=0x7fffea882470) at
>> > auth/cephx/../Crypto.h:110
>> > #1  encode_encrypt_enc_bl<CephXChallengeBlob> (cct=<optimized out>,
>> > error="", out=..., key=..., t=<synthetic pointer>) at
>> > auth/cephx/CephxProtocol.h:464
>> > #2  encode_encrypt<CephXChallengeBlob> (cct=<optimized out>,
>> > error="",
>> > out=..., key=..., t=<synthetic pointer>) at
>> > auth/cephx/CephxProtocol.h:489
>> > #3  cephx_calc_client_server_challenge (cct=<optimized out>,
>> > secret=..., server_challenge=9622349603176979543,
>> > client_challenge=7732813711656640623, key=key@entry=0x7fffea8824a8,
>> > ret="")
>> >     at auth/cephx/CephxProtocol.cc:36
>> > #4  0x00007ffff313aff4 in CephxClientHandler::build_request
>> > (this=0x7fffd4001520, bl=...) at
>> auth/cephx/CephxClientHandler.cc:53
>> > #5  0x00007ffff2fe4a79 in MonClient::handle_auth (this=this@entry=0
>> x7
>> > ff
>> > fec006b70, m=m@entry=0x7fffd0002ee0) at mon/MonClient.cc:510
>> > #6  0x00007ffff2fe6507 in MonClient::ms_dispatch
>> > (this=0x7fffec006b70,
>> > m=0x7fffd0002ee0) at mon/MonClient.cc:277
>> > #7  0x00007ffff30d5dc9 in ms_deliver_dispatch (m=0x7fffd0002ee0,
>> > this=0x7fffec055410) at ./msg/Messenger.h:582
>> > #8  DispatchQueue::entry (this=0x7fffec0555d8) at
>> > msg/simple/DispatchQueue.cc:185
>> > #9  0x00007ffff31023bd in DispatchQueue::DispatchThread::entry
>> > (this=<optimized out>) at msg/simple/DispatchQueue.h:103
>> > #10 0x00007ffff7bc4182 in start_thread () from /lib/x86_64-linux-
>> > gnu/libpthread.so.0
>> > #11 0x00007ffff78f147d in clone () from /lib/x86_64-linux-
>> > gnu/libc.so.6
>> >
>> > -----Original Message-----
>> > From: John Spray <jspray@xxxxxxxxxx>
>> > To: Mathias Buresch <mathias.buresch@xxxxxxxxxxxx>
>> > Cc: ceph-users@xxxxxxxx <ceph-users@xxxxxxxx>
>> > Subject: Re:  Ceph Status - Segmentation Fault
>> > Date: Wed, 25 May 2016 15:41:51 +0100
>> >
>> > On Wed, May 25, 2016 at 3:00 PM, Mathias Buresch
>> > <mathias.buresch@xxxxxxxxxxxx> wrote:
>> > >
>> > >
>> > > I don't know what exactly is segfaulting.
>> > >
>> > > Here ist the output with command line flags and gdb (I can't
>> really
>> > > notice erros in that output):
>> > >
>> > > # ceph -s --debug-monc=20 --debug-ms=20
>> > > 2016-05-25 14:51:02.406135 7f188300a700 10 monclient(hunting):
>> > > build_initial_monmap
>> > > 2016-05-25 14:51:02.406444 7f188300a700 10 -- :/0 ready :/0
>> > > 2016-05-25 14:51:02.407214 7f188300a700  1 -- :/0 messenger.start
>> > > 2016-05-25 14:51:02.407261 7f188300a700 10 monclient(hunting):
>> init
>> > > 2016-05-25 14:51:02.407291 7f188300a700 10 monclient(hunting):
>> > > auth_supported 2 method cephx
>> > > 2016-05-25 14:51:02.407312 7f187b7fe700 10 -- :/2987460054
>> > > reaper_entry
>> > > start
>> > > 2016-05-25 14:51:02.407380 7f187b7fe700 10 -- :/2987460054 reaper
>> > > 2016-05-25 14:51:02.407383 7f187b7fe700 10 -- :/2987460054 reaper
>> > > done
>> > > 2016-05-25 14:51:02.407638 7f188300a700 10 monclient(hunting):
>> > > _reopen_session rank -1 name
>> > > 2016-05-25 14:51:02.407646 7f188300a700 10 -- :/2987460054
>> > > connect_rank
>> > > to 62.176.141.181:6789/0, creating pipe and registering
>> > > 2016-05-25 14:51:02.407686 7f188300a700 10 -- :/2987460054 >>
>> > > 62.176.141.181:6789/0 pipe(0x7f187c064010 sd=-1 :0 s=1 pgs=0 cs=0
>> > > l=1
>> > > c=0x7f187c05aa50).register_pipe
>> > > 2016-05-25 14:51:02.407698 7f188300a700 10 -- :/2987460054
>> > > get_connection mon.0 62.176.141.181:6789/0 new 0x7f187c064010
>> > > 2016-05-25 14:51:02.407693 7f1879ffb700 10 -- :/2987460054 >>
>> > > 62.176.141.181:6789/0 pipe(0x7f187c064010 sd=-1 :0 s=1 pgs=0 cs=0
>> > > l=1
>> > > c=0x7f187c05aa50).writer: state = connecting policy.server=0
>> > > 2016-05-25 14:51:02.407723 7f1879ffb700 10 -- :/2987460054 >>
>> > > 62.176.141.181:6789/0 pipe(0x7f187c064010 sd=-1 :0 s=1 pgs=0 cs=0
>> > > l=1
>> > > c=0x7f187c05aa50).connect 0
>> > > 2016-05-25 14:51:02.407738 7f188300a700 10 monclient(hunting):
>> > > picked
>> > > mon.pix01 con 0x7f187c05aa50 addr 62.176.141.181:6789/0
>> > > 2016-05-25 14:51:02.407745 7f188300a700 20 -- :/2987460054
>> > > send_keepalive con 0x7f187c05aa50, have pipe.
>> > > 2016-05-25 14:51:02.407744 7f1879ffb700 10 -- :/2987460054 >>
>> > > 62.176.141.181:6789/0 pipe(0x7f187c064010 sd=3 :0 s=1 pgs=0 cs=0
>> > > l=1
>> > > c=0x7f187c05aa50).connecting to 62.176.141.181:6789/0
>> > > 2016-05-25 14:51:02.407759 7f188300a700 10 monclient(hunting):
>> > > _send_mon_message to mon.pix01 at 62.176.141.181:6789/0
>> > > 2016-05-25 14:51:02.407763 7f188300a700  1 -- :/2987460054 -->
>> > > 62.176.141.181:6789/0 -- auth(proto 0 30 bytes epoch 0) v1 -- ?+0
>> > > 0x7f187c060380 con 0x7f187c05aa50
>> > > 2016-05-25 14:51:02.407768 7f188300a700 20 -- :/2987460054
>> > > submit_message auth(proto 0 30 bytes epoch 0) v1 remote,
>> > > 62.176.141.181:6789/0, have pipe.
>> > > 2016-05-25 14:51:02.407773 7f188300a700 10 monclient(hunting):
>> > > renew_subs
>> > > 2016-05-25 14:51:02.407777 7f188300a700 10 monclient(hunting):
>> > > authenticate will time out at 2016-05-25 14:56:02.407777
>> > > 2016-05-25 14:51:02.408128 7f1879ffb700 20 -- :/2987460054 >>
>> > > 62.176.141.181:6789/0 pipe(0x7f187c064010 sd=3 :37964 s=1 pgs=0
>> > > cs=0
>> > > l=1 c=0x7f187c05aa50).connect read peer addr
>> 62.176.141.181:6789/0
>> > > on
>> > > socket 3
>> > > 2016-05-25 14:51:02.408144 7f1879ffb700 20 -- :/2987460054 >>
>> > > 62.176.141.181:6789/0 pipe(0x7f187c064010 sd=3 :37964 s=1 pgs=0
>> > > cs=0
>> > > l=1 c=0x7f187c05aa50).connect peer addr for me is
>> > > 62.176.141.181:37964/0
>> > > 2016-05-25 14:51:02.408148 7f1879ffb700  1 --
>> > > 62.176.141.181:0/2987460054 learned my addr
>> > > 62.176.141.181:0/2987460054
>> > > 2016-05-25 14:51:02.408188 7f1879ffb700 10 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=1 pgs=0 cs=0 l=1
>> > > c=0x7f187c05aa50).connect sent my addr
>> 62.176.141.181:0/2987460054
>> > > 2016-05-25 14:51:02.408197 7f1879ffb700 10 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=1 pgs=0 cs=0 l=1
>> > > c=0x7f187c05aa50).connect sending gseq=1 cseq=0 proto=15
>> > > 2016-05-25 14:51:02.408207 7f1879ffb700 20 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=1 pgs=0 cs=0 l=1
>> > > c=0x7f187c05aa50).connect wrote (self +) cseq, waiting for reply
>> > > 2016-05-25 14:51:02.408259 7f1879ffb700 20 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=1 pgs=0 cs=0 l=1
>> > > c=0x7f187c05aa50).connect got reply tag 1 connect_seq 1
>> global_seq
>> > > 327710 proto 15 flags 1 features 55169095435288575
>> > > 2016-05-25 14:51:02.408269 7f1879ffb700 10 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).connect success 1, lossy = 1, features
>> > > 55169095435288575
>> > > 2016-05-25 14:51:02.408280 7f1879ffb700 20 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).connect starting reader
>> > > 2016-05-25 14:51:02.408325 7f1879ffb700 10 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).writer: state = open policy.server=0
>> > > 2016-05-25 14:51:02.408343 7f1879ffb700 10 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).write_keepalive2 14 2016-05-25 14:51:02.408342
>> > > 2016-05-25 14:51:02.408378 7f1879ffb700 20 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).writer encoding 1 features 55169095435288575
>> > > 0x7f187c060380 auth(proto 0 30 bytes epoch 0) v1
>> > > 2016-05-25 14:51:02.408356 7f1879efa700 20 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).reader reading tag...
>> > > 2016-05-25 14:51:02.408406 7f1879ffb700 20 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).writer no session security
>> > > 2016-05-25 14:51:02.408415 7f1879ffb700 20 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).writer sending 1 0x7f187c060380
>> > > 2016-05-25 14:51:02.408453 7f1879ffb700 10 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).writer: state = open policy.server=0
>> > > 2016-05-25 14:51:02.408455 7f1879efa700 20 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).reader got KEEPALIVE_ACK
>> > > 2016-05-25 14:51:02.408463 7f1879ffb700 20 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).writer sleeping
>> > > 2016-05-25 14:51:02.408482 7f1879efa700 20 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).reader reading tag...
>> > > 2016-05-25 14:51:02.408696 7f1879efa700 20 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).reader got ACK
>> > > 2016-05-25 14:51:02.408713 7f1879efa700 15 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).reader got ack seq 1
>> > > 2016-05-25 14:51:02.408721 7f1879efa700 20 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).reader reading tag...
>> > > 2016-05-25 14:51:02.408732 7f1879efa700 20 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).reader got MSG
>> > > 2016-05-25 14:51:02.408739 7f1879efa700 20 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).reader got envelope type=4 src mon.0 front=340
>> > > data=0
>> > > off 0
>> > > 2016-05-25 14:51:02.408751 7f1879efa700 10 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).reader wants 340 from dispatch throttler
>> > > 0/104857600
>> > > 2016-05-25 14:51:02.408763 7f1879efa700 20 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).reader got front 340
>> > > 2016-05-25 14:51:02.408770 7f1879efa700 10 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).aborted = 0
>> > > 2016-05-25 14:51:02.408776 7f1879efa700 20 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).reader got 340 + 0 + 0 byte message
>> > > 2016-05-25 14:51:02.408801 7f1879efa700 10 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).No session security set
>> > > 2016-05-25 14:51:02.408813 7f1879efa700 10 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).reader got message 1 0x7f186c001cb0 mon_map
>> > > magic:
>> > > 0
>> > > v1
>> > > 2016-05-25 14:51:02.408827 7f1879efa700 20 --
>> > > 62.176.141.181:0/2987460054 queue 0x7f186c001cb0 prio 196
>> > > 2016-05-25 14:51:02.408837 7f1879efa700 20 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).reader reading tag...
>> > > 2016-05-25 14:51:02.408851 7f1879ffb700 10 --
>> > > 62.176.141.181:0/2987460054 >> 62.176.141.181:6789/0
>> > > pipe(0x7f187c064010 sd=3 :37964 s=2 pgs=327710 cs=1 l=1
>> > > c=0x7f187c05aa50).writer: state = open policy.server=0
>> > > Segmentation fault
>> > >
>> > >
>> > > (gdb) run /usr/bin/ceph status
>> > > Starting program: /usr/bin/python /usr/bin/ceph status
>> > > [Thread debugging using libthread_db enabled]
>> > > Using host libthread_db library "/lib/x86_64-linux-
>> > > gnu/libthread_db.so.1".
>> > > [New Thread 0x7ffff10f5700 (LWP 23401)]
>> > > [New Thread 0x7ffff08f4700 (LWP 23402)]
>> > > [Thread 0x7ffff10f5700 (LWP 23401) exited]
>> > > [New Thread 0x7ffff10f5700 (LWP 23403)]
>> > > [Thread 0x7ffff10f5700 (LWP 23403) exited]
>> > > [New Thread 0x7ffff10f5700 (LWP 23404)]
>> > > [Thread 0x7ffff10f5700 (LWP 23404) exited]
>> > > [New Thread 0x7ffff10f5700 (LWP 23405)]
>> > > [Thread 0x7ffff10f5700 (LWP 23405) exited]
>> > > [New Thread 0x7ffff10f5700 (LWP 23406)]
>> > > [Thread 0x7ffff10f5700 (LWP 23406) exited]
>> > > [New Thread 0x7ffff10f5700 (LWP 23407)]
>> > > [Thread 0x7ffff10f5700 (LWP 23407) exited]
>> > > [New Thread 0x7ffff10f5700 (LWP 23408)]
>> > > [New Thread 0x7fffeb885700 (LWP 23409)]
>> > > [New Thread 0x7fffeb084700 (LWP 23410)]
>> > > [New Thread 0x7fffea883700 (LWP 23411)]
>> > > [New Thread 0x7fffea082700 (LWP 23412)]
>> > > [New Thread 0x7fffe9881700 (LWP 23413)]
>> > > [New Thread 0x7fffe9080700 (LWP 23414)]
>> > > [New Thread 0x7fffe887f700 (LWP 23415)]
>> > > [New Thread 0x7fffe807e700 (LWP 23416)]
>> > > [New Thread 0x7fffe7f7d700 (LWP 23419)]
>> > >
>> > > Program received signal SIGSEGV, Segmentation fault.
>> > > [Switching to Thread 0x7fffea883700 (LWP 23411)]
>> > > 0x00007ffff3141a57 in ?? () from /usr/lib/librados.so.2
>> > > (gdb) bt
>> > > #0  0x00007ffff3141a57 in ?? () from /usr/lib/librados.so.2
>> > > #1  0x00007ffff313aff4 in ?? () from /usr/lib/librados.so.2
>> > > #2  0x00007ffff2fe4a79 in ?? () from /usr/lib/librados.so.2
>> > > #3  0x00007ffff2fe6507 in ?? () from /usr/lib/librados.so.2
>> > > #4  0x00007ffff30d5dc9 in ?? () from /usr/lib/librados.so.2
>> > > #5  0x00007ffff31023bd in ?? () from /usr/lib/librados.so.2
>> > > #6  0x00007ffff7bc4182 in start_thread () from /lib/x86_64-linux-
>> > > gnu/libpthread.so.0
>> > > #7  0x00007ffff78f147d in clone () from /lib/x86_64-linux-
>> > > gnu/libc.so.6
>> > >
>> > >
>> > > Does that help? I cant really see where the error is. :)
>> > Hmm, can you try getting that backtrace again after installing the
>> > ceph-debuginfo package?  Also add --debug-rados=20 to your command
>> > line (you can use all the --debug... options when you're running
>> > inside gdb to get the logs and the backtrace in one).
>> >
>> > John
>> >
>> > >
>> > >
>> > >
>> > > -----Original Message-----
>> > > From: John Spray <jspray@xxxxxxxxxx>
>> > > To: Mathias Buresch <mathias.buresch@xxxxxxxxxxxx>
>> > > Cc: ceph-users@xxxxxxxx <ceph-users@xxxxxxxx>
>> > > Subject: Re:  Ceph Status - Segmentation Fault
>> > > Date: Wed, 25 May 2016 10:16:55 +0100
>> > >
>> > > On Mon, May 23, 2016 at 12:41 PM, Mathias Buresch
>> > > <mathias.buresch@xxxxxxxxxxxx> wrote:
>> > > >
>> > > >
>> > > >
>> > > > Please found the logs with higher debug level attached to this
>> > > > email.
>> > > You've attached the log from your mon, but it's not your mon
>> that's
>> > > segfaulting, right?
>> > >
>> > > You can use normal ceph command line flags to crank up the
>> > > verbosity
>> > > on the CLI too (--debug-monc=20 --debug-ms=20 spring to mind).
>> > >
>> > > You can also run the ceph CLI in gdb like this:
>> > > gdb python
>> > > (gdb) run /usr/bin/ceph status
>> > > ... hopefully it crashes and then ...
>> > > (gdb) bt
>> > >
>> > > Cheers,
>> > > John
>> > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > > Kind regards
>> > > > Mathias
>> > > > _______________________________________________
>> > > > ceph-users mailing list
>> > > > ceph-users@xxxxxxxxxxxxxx
>> > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> > > >
>> > _______________________________________________
>> > ceph-users mailing list
>> > ceph-users@xxxxxxxxxxxxxx
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >
>>



-- 
Cheers,
Brad
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux