Re: GlusterFS 3.7.11 crash issue

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

gdb doesn't accept the coredump file ...

root@nfs05 /var/crash # gdb /usr/sbin/glusterfsd /var/crash/_usr_sbin_glusterfsd.0.crash
GNU gdb (Ubuntu 7.7.1-0ubuntu5~14.04.2) 7.7.1
Copyright (C) 2014 Free Software Foundation, Inc.
License GPLv3+: GNU GPL version 3 or later <http://gnu.org/licenses/gpl.html>
This is free software: you are free to change and redistribute it.
There is NO WARRANTY, to the extent permitted by law.  Type "show copying"
and "show warranty" for details.
This GDB was configured as "x86_64-linux-gnu".
Type "show configuration" for configuration details.
For bug reporting instructions, please see:
<http://www.gnu.org/software/gdb/bugs/>.
Find the GDB manual and other documentation resources online at:
<http://www.gnu.org/software/gdb/documentation/>.
For help, type "help".
Type "apropos word" to search for commands related to "word"...
Reading symbols from /usr/sbin/glusterfsd...(no debugging symbols found)...done.
"/var/crash/_usr_sbin_glusterfsd.0.crash" is not a core dump: File format not recognized
(gdb) bt
No stack.

Yann


Le 29/06/2016 08:39, Anoop C S a écrit :
On Tue, 2016-06-28 at 10:49 +0200, Yann LEMARIE wrote:
Hi,

I found the coredump file, but it's a 15Mo file (zipped), I can't
post it on this mailling list.

Great. In order to exactly pin point the crash location, can you please
attach gdb to extracted coredump file and share us the complete back
trace by executing `bt` command in gdb shell? Apart from gdb you may be
instructed to install some debug-info packages for extracting a useful
back trace while attaching gdb as follows:

# gdb /usr/sbin/glusterfsd <path-to-coredump-file>

If prompted install required packages and reattach the coredump file.
When you are inside (gdb) prompt type 'bt' and paste the back trace.

Here is some parts of the repport :

ProblemType: Crash
Architecture: amd64
Date: Sun Jun 26 11:27:44 2016
DistroRelease: Ubuntu 14.04
ExecutablePath: /usr/sbin/glusterfsd
ExecutableTimestamp: 1460982898
ProcCmdline: /usr/sbin/glusterfsd -s nfs05 --volfile-id
cdn.nfs05.srv-cdn -p /var/lib/glusterd/vols/cdn/run/nfs05-srv-
cdn.pid -S /var/run/gluster/d52ac3e6c0a3fa316a9e8360976f3af5.socket
--brick-name /srv/cdn -l /var/log/glusterfs/bricks/srv-cdn.log --
xlator-option *-posix.glusterd-uuid=6af63b78-a3da-459d-a909-
c010e6c9072c --brick-port 49155 --xlator-option cdn-server.listen-
port=49155
ProcCwd: /
ProcEnviron:
 PATH=(custom, no user)
 TERM=linux
ProcMaps:
 7f25f18d9000-7f25f18da000 ---p 00000000 00:00 0
 7f25f18da000-7f25f19da000 rw-p 00000000 00:00
0                          [stack:849]
 7f25f19da000-7f25f19db000 ---p 00000000 00:00 0
 ...
ProcStatus:
 Name:  glusterfsd
 State: D (disk sleep)
 Tgid:  7879
 Ngid:  0
 Pid:   7879
 PPid:  1
 TracerPid:     0
 Uid:   0       0       0       0
 Gid:   0       0       0       0
 FDSize:        64
 Groups:        0
 VmPeak:          878404 kB
 VmSize:          878404 kB
 VmLck:        0 kB
 VmPin:        0 kB
 VmHWM:    96104 kB
 VmRSS:    90652 kB
 VmData:          792012 kB
 VmStk:      276 kB
 VmExe:       84 kB
 VmLib:     7716 kB
 VmPTE:      700 kB
 VmSwap:           20688 kB
 Threads:       22
 SigQ:  0/30034
 SigPnd:        0000000000000000
 ShdPnd:        0000000000000000
 SigBlk:        0000000000004a01
 SigIgn:        0000000000001000
 SigCgt:        00000001800000fa
 CapInh:        0000000000000000
 CapPrm:        0000001fffffffff
 CapEff:        0000001fffffffff
 CapBnd:        0000001fffffffff
 Seccomp:       0
 Cpus_allowed:  7fff
 Cpus_allowed_list:     0-14
 Mems_allowed:  00000000,00000001
 Mems_allowed_list:     0
 voluntary_ctxt_switches:       3
 nonvoluntary_ctxt_switches:    1
Signal: 11
Uname: Linux 3.13.0-44-generic x86_64
UserGroups:
CoreDump: base64
 ...

Yann

Le 28/06/2016 09:31, Anoop C S a écrit :
On Mon, 2016-06-27 at 15:05 +0200, Yann LEMARIE wrote:
 @Anoop,

Where can I find the coredump file ?

You will get hints about the crash from entries inside
/var/log/messages(for example pid of the process, location of
coredump
etc). 

The crash occurs 2 times last 7 days, each time a sunday morning
with
no reason, no increase of traffic or something like this, the
volume
was mounted since 15 days.

The bricks are used as a CDN like, distributting small images and
css
files with a nginx https service (with a load balancer and 2
EC2), on
a sunday morning there is not a lot of activity ...

From the very minimal back trace that we have from brick logs I
would
assume that a truncate operation was being handled by trash
translator
and it crashed.

Volume infos: 
root@nfs05 /var/log/glusterfs # gluster volume info cdn
 
Volume Name: cdn
Type: Replicate
Volume ID: c53b9bae-5e12-4f13-8217-53d8c96c302c
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: nfs05:/srv/cdn
Brick2: nfs06:/srv/cdn
Options Reconfigured:
performance.readdir-ahead: on
features.trash: on
features.trash-max-filesize: 20MB
 
I don't know if there is a link with this crash problem, but I
have
another problem with my 2 servers that make GluserFS's clients
disconnected (from another volume) :
Jun 24 02:28:04 nfs05 kernel: [2039468.818617] xen_netfront:
xennet: skb rides the rocket: 19 slots
Jun 24 02:28:11 nfs05 kernel: [2039475.744086] net_ratelimit:
66
callbacks suppressed
 It seem to be a network interface problem :
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1317811

Yann

Le 27/06/2016 12:59, Anoop C S a écrit :
On Mon, 2016-06-27 at 09:47 +0200, Yann LEMARIE wrote:
Hi,

I'm using GlusterFS since many years and never see this
problem,
but
this is the second time in one week ...

I have 3 volumes with 2 bricks and 1 volume crash with no
reason,
Did you observe the crash while mounting the volume? Or can you
be
more
specific on what were you doing just before you saw the crash?
Can
you
please share the output of `gluster volume info <VOLNAME>`?

 I just have to stop/start the volume to make it up again.
The only logs I can find are in syslog :

Jun 26 11:27:44 nfs05 srv-cdn[7879]: pending frames:
Jun 26 11:27:44 nfs05 srv-cdn[7879]: frame : type(0) op(10)
Jun 26 11:27:44 nfs05 srv-cdn[7879]: patchset:
git://git.gluster.com/glusterfs.git
Jun 26 11:27:44 nfs05 srv-cdn[7879]: signal received: 11
Jun 26 11:27:44 nfs05 srv-cdn[7879]: time of crash:
Jun 26 11:27:44 nfs05 srv-cdn[7879]: 2016-06-26 09:27:44
Jun 26 11:27:44 nfs05 srv-cdn[7879]: configuration details:
Jun 26 11:27:44 nfs05 srv-cdn[7879]: argp 1
Jun 26 11:27:44 nfs05 srv-cdn[7879]: backtrace 1
Jun 26 11:27:44 nfs05 srv-cdn[7879]: dlfcn 1
Jun 26 11:27:44 nfs05 srv-cdn[7879]: libpthread 1
Jun 26 11:27:44 nfs05 srv-cdn[7879]: llistxattr 1
Jun 26 11:27:44 nfs05 srv-cdn[7879]: setfsid 1
Jun 26 11:27:44 nfs05 srv-cdn[7879]: spinlock 1
Jun 26 11:27:44 nfs05 srv-cdn[7879]: epoll.h 1
Jun 26 11:27:44 nfs05 srv-cdn[7879]: xattr.h 1
Jun 26 11:27:44 nfs05 srv-cdn[7879]: st_atim.tv_nsec 1
Jun 26 11:27:44 nfs05 srv-cdn[7879]: package-string:
glusterfs
3.7.11
Jun 26 11:27:44 nfs05 srv-cdn[7879]: ---------

Jun 26 11:27:44 nfs06 srv-cdn[1787]: pending frames:
Jun 26 11:27:44 nfs06 srv-cdn[1787]: frame : type(0) op(10)
Jun 26 11:27:44 nfs06 srv-cdn[1787]: patchset:
git://git.gluster.com/glusterfs.git
Jun 26 11:27:44 nfs06 srv-cdn[1787]: signal received: 11
Jun 26 11:27:44 nfs06 srv-cdn[1787]: time of crash:
Jun 26 11:27:44 nfs06 srv-cdn[1787]: 2016-06-26 09:27:44
Jun 26 11:27:44 nfs06 srv-cdn[1787]: configuration details:
Jun 26 11:27:44 nfs06 srv-cdn[1787]: argp 1
Jun 26 11:27:44 nfs06 srv-cdn[1787]: backtrace 1
Jun 26 11:27:44 nfs06 srv-cdn[1787]: dlfcn 1
Jun 26 11:27:44 nfs06 srv-cdn[1787]: libpthread 1
Jun 26 11:27:44 nfs06 srv-cdn[1787]: llistxattr 1
Jun 26 11:27:44 nfs06 srv-cdn[1787]: setfsid 1
Jun 26 11:27:44 nfs06 srv-cdn[1787]: spinlock 1
Jun 26 11:27:44 nfs06 srv-cdn[1787]: epoll.h 1
Jun 26 11:27:44 nfs06 srv-cdn[1787]: xattr.h 1
Jun 26 11:27:44 nfs06 srv-cdn[1787]: st_atim.tv_nsec 1
Jun 26 11:27:44 nfs06 srv-cdn[1787]: package-string:
glusterfs
3.7.11
Jun 26 11:27:44 nfs06 srv-cdn[1787]: ---------

 
Thanks for your help


Regards
-- 
Yann Lemarié
iRaiser - Support Technique
 
ylemarie@xxxxxxxxxx
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users
 
-- 
Yann Lemarié
iRaiser - Support Technique
 
ylemarie@xxxxxxxxxx


_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users
 
-- 
Yann Lemarié
iRaiser - Support Technique
 
ylemarie@xxxxxxxxxx
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users

--
Yann Lemarié
iRaiser - Support Technique
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux