Hi, gdb doesn't accept the coredump file ... root@nfs05 /var/crash # gdb /usr/sbin/glusterfsd /var/crash/_usr_sbin_glusterfsd.0.crash Yann Le 29/06/2016 08:39, Anoop C S a
écrit :
On Tue, 2016-06-28 at 10:49 +0200, Yann LEMARIE wrote:Hi, I found the coredump file, but it's a 15Mo file (zipped), I can't post it on this mailling list.Great. In order to exactly pin point the crash location, can you please attach gdb to extracted coredump file and share us the complete back trace by executing `bt` command in gdb shell? Apart from gdb you may be instructed to install some debug-info packages for extracting a useful back trace while attaching gdb as follows: # gdb /usr/sbin/glusterfsd <path-to-coredump-file> If prompted install required packages and reattach the coredump file. When you are inside (gdb) prompt type 'bt' and paste the back trace.Here is some parts of the repport :ProblemType: Crash Architecture: amd64 Date: Sun Jun 26 11:27:44 2016 DistroRelease: Ubuntu 14.04 ExecutablePath: /usr/sbin/glusterfsd ExecutableTimestamp: 1460982898 ProcCmdline: /usr/sbin/glusterfsd -s nfs05 --volfile-id cdn.nfs05.srv-cdn -p /var/lib/glusterd/vols/cdn/run/nfs05-srv- cdn.pid -S /var/run/gluster/d52ac3e6c0a3fa316a9e8360976f3af5.socket --brick-name /srv/cdn -l /var/log/glusterfs/bricks/srv-cdn.log -- xlator-option *-posix.glusterd-uuid=6af63b78-a3da-459d-a909- c010e6c9072c --brick-port 49155 --xlator-option cdn-server.listen- port=49155 ProcCwd: / ProcEnviron: PATH=(custom, no user) TERM=linux ProcMaps: 7f25f18d9000-7f25f18da000 ---p 00000000 00:00 0 7f25f18da000-7f25f19da000 rw-p 00000000 00:00 0 [stack:849] 7f25f19da000-7f25f19db000 ---p 00000000 00:00 0...ProcStatus: Name: glusterfsd State: D (disk sleep) Tgid: 7879 Ngid: 0 Pid: 7879 PPid: 1 TracerPid: 0 Uid: 0 0 0 0 Gid: 0 0 0 0 FDSize: 64 Groups: 0 VmPeak: 878404 kB VmSize: 878404 kB VmLck: 0 kB VmPin: 0 kB VmHWM: 96104 kB VmRSS: 90652 kB VmData: 792012 kB VmStk: 276 kB VmExe: 84 kB VmLib: 7716 kB VmPTE: 700 kB VmSwap: 20688 kB Threads: 22 SigQ: 0/30034 SigPnd: 0000000000000000 ShdPnd: 0000000000000000 SigBlk: 0000000000004a01 SigIgn: 0000000000001000 SigCgt: 00000001800000fa CapInh: 0000000000000000 CapPrm: 0000001fffffffff CapEff: 0000001fffffffff CapBnd: 0000001fffffffff Seccomp: 0 Cpus_allowed: 7fff Cpus_allowed_list: 0-14 Mems_allowed: 00000000,00000001 Mems_allowed_list: 0 voluntary_ctxt_switches: 3 nonvoluntary_ctxt_switches: 1 Signal: 11 Uname: Linux 3.13.0-44-generic x86_64 UserGroups: CoreDump: base64... Yann Le 28/06/2016 09:31, Anoop C S a écrit :On Mon, 2016-06-27 at 15:05 +0200, Yann LEMARIE wrote:@Anoop, Where can I find the coredump file ?You will get hints about the crash from entries inside /var/log/messages(for example pid of the process, location of coredump etc).The crash occurs 2 times last 7 days, each time a sunday morning with no reason, no increase of traffic or something like this, the volume was mounted since 15 days. The bricks are used as a CDN like, distributting small images and css files with a nginx https service (with a load balancer and 2 EC2), on a sunday morning there is not a lot of activity ...From the very minimal back trace that we have from brick logs I would assume that a truncate operation was being handled by trash translator and it crashed.Volume infos:root@nfs05 /var/log/glusterfs # gluster volume info cdn Volume Name: cdn Type: Replicate Volume ID: c53b9bae-5e12-4f13-8217-53d8c96c302c Status: Started Number of Bricks: 1 x 2 = 2 Transport-type: tcp Bricks: Brick1: nfs05:/srv/cdn Brick2: nfs06:/srv/cdn Options Reconfigured: performance.readdir-ahead: on features.trash: on features.trash-max-filesize: 20MBI don't know if there is a link with this crash problem, but I have another problem with my 2 servers that make GluserFS's clients disconnected (from another volume) :Jun 24 02:28:04 nfs05 kernel: [2039468.818617] xen_netfront: xennet: skb rides the rocket: 19 slots Jun 24 02:28:11 nfs05 kernel: [2039475.744086] net_ratelimit: 66 callbacks suppressedIt seem to be a network interface problem : https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1317811 Yann Le 27/06/2016 12:59, Anoop C S a écrit :On Mon, 2016-06-27 at 09:47 +0200, Yann LEMARIE wrote:Hi, I'm using GlusterFS since many years and never see this problem, but this is the second time in one week ... I have 3 volumes with 2 bricks and 1 volume crash with no reason,Did you observe the crash while mounting the volume? Or can you be more specific on what were you doing just before you saw the crash? Can you please share the output of `gluster volume info <VOLNAME>`?I just have to stop/start the volume to make it up again. The only logs I can find are in syslog :Jun 26 11:27:44 nfs05 srv-cdn[7879]: pending frames: Jun 26 11:27:44 nfs05 srv-cdn[7879]: frame : type(0) op(10) Jun 26 11:27:44 nfs05 srv-cdn[7879]: patchset: git://git.gluster.com/glusterfs.git Jun 26 11:27:44 nfs05 srv-cdn[7879]: signal received: 11 Jun 26 11:27:44 nfs05 srv-cdn[7879]: time of crash: Jun 26 11:27:44 nfs05 srv-cdn[7879]: 2016-06-26 09:27:44 Jun 26 11:27:44 nfs05 srv-cdn[7879]: configuration details: Jun 26 11:27:44 nfs05 srv-cdn[7879]: argp 1 Jun 26 11:27:44 nfs05 srv-cdn[7879]: backtrace 1 Jun 26 11:27:44 nfs05 srv-cdn[7879]: dlfcn 1 Jun 26 11:27:44 nfs05 srv-cdn[7879]: libpthread 1 Jun 26 11:27:44 nfs05 srv-cdn[7879]: llistxattr 1 Jun 26 11:27:44 nfs05 srv-cdn[7879]: setfsid 1 Jun 26 11:27:44 nfs05 srv-cdn[7879]: spinlock 1 Jun 26 11:27:44 nfs05 srv-cdn[7879]: epoll.h 1 Jun 26 11:27:44 nfs05 srv-cdn[7879]: xattr.h 1 Jun 26 11:27:44 nfs05 srv-cdn[7879]: st_atim.tv_nsec 1 Jun 26 11:27:44 nfs05 srv-cdn[7879]: package-string: glusterfs 3.7.11 Jun 26 11:27:44 nfs05 srv-cdn[7879]: --------- Jun 26 11:27:44 nfs06 srv-cdn[1787]: pending frames: Jun 26 11:27:44 nfs06 srv-cdn[1787]: frame : type(0) op(10) Jun 26 11:27:44 nfs06 srv-cdn[1787]: patchset: git://git.gluster.com/glusterfs.git Jun 26 11:27:44 nfs06 srv-cdn[1787]: signal received: 11 Jun 26 11:27:44 nfs06 srv-cdn[1787]: time of crash: Jun 26 11:27:44 nfs06 srv-cdn[1787]: 2016-06-26 09:27:44 Jun 26 11:27:44 nfs06 srv-cdn[1787]: configuration details: Jun 26 11:27:44 nfs06 srv-cdn[1787]: argp 1 Jun 26 11:27:44 nfs06 srv-cdn[1787]: backtrace 1 Jun 26 11:27:44 nfs06 srv-cdn[1787]: dlfcn 1 Jun 26 11:27:44 nfs06 srv-cdn[1787]: libpthread 1 Jun 26 11:27:44 nfs06 srv-cdn[1787]: llistxattr 1 Jun 26 11:27:44 nfs06 srv-cdn[1787]: setfsid 1 Jun 26 11:27:44 nfs06 srv-cdn[1787]: spinlock 1 Jun 26 11:27:44 nfs06 srv-cdn[1787]: epoll.h 1 Jun 26 11:27:44 nfs06 srv-cdn[1787]: xattr.h 1 Jun 26 11:27:44 nfs06 srv-cdn[1787]: st_atim.tv_nsec 1 Jun 26 11:27:44 nfs06 srv-cdn[1787]: package-string: glusterfs 3.7.11 Jun 26 11:27:44 nfs06 srv-cdn[1787]: ---------Thanks for your help Regards -- Yann Lemarié iRaiser - Support Technique ylemarie@xxxxxxxxxx _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users-- Yann Lemarié iRaiser - Support Technique ylemarie@xxxxxxxxxx _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users-- Yann Lemarié iRaiser - Support Technique ylemarie@xxxxxxxxxx _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users --
|
_______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users