Re: Kernel oops when GFS2 used as localfs

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Steven Whitehouse wrote:
Hi,

This looks like Red Hat bugzilla 229831 and if so then there is already
a fix in Linus' upstream kernel and in the latest kernel build for
Fedora (both 5 and 6),

yeah... agree. I missed the unlink part of the trace in previous post and thought it might have something to do with my ino number hacking in the lookup code.

-- Wendy
Steve.

On Wed, 2007-03-14 at 00:15 -0700, Sridhar Ramaswamy (srramasw) wrote:
Sure Wendy. Here it is,

"fs/gfs2/inode.c" 1256
struct inode *gfs2_ilookup(struct super_block *sb, struct gfs2_inum_host
*inum)
{
        return ilookup5(sb, (unsigned long)inum->no_formal_ino,
                        iget_test, inum);
}

static struct inode *gfs2_iget(struct super_block *sb, struct
gfs2_inum_host *inum)
{
        return iget5_locked(sb, (unsigned long)inum->no_formal_ino,
                     iget_test, iget_set, inum);
}

BTW - this code is from kernel version 2.6.20.1.
Thanks,
Sridhar

-----Original Message-----
From: linux-cluster-bounces@xxxxxxxxxx [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Wendy Cheng
Sent: Tuesday, March 13, 2007 10:29 PM
To: linux clustering
Subject: Re:  Kernel oops when GFS2 used as localfs

Sridhar Ramaswamy (srramasw) wrote:

Mar 13 15:25:56 cfs1 kernel: ------------[ cut here ]------------
Mar 13 15:25:56 cfs1 kernel: kernel BUG at fs/gfs2/meta_io.c:474!
I don't have time to pull kernel.org source code right now. Could you cut-and-paste the following two routines (from fs/gfs2/inode.c): gfs2_ilookup() and gfs2_iget() so we can look into this quickly ? They are all one-liner so shouldn't be too much troubles. GFS2 currently has ino issue with its lookup code.

-- Wendy

Mar 13 15:25:56 cfs1 kernel: invalid opcode: 0000 [#1]
Mar 13 15:25:56 cfs1 kernel: SMP
Mar 13 15:25:56 cfs1 kernel: Modules linked in: lock_nolock gfs2 reiserfs nfsd exportfs nfs lockd nfs_acl ipv6 parport_pc lp parport autofs4 sunrpc dm_mirror dm_mod button battery ac uhci_hcd ehci_hcd intel_rng rng_core i2c_i801 i2c_core e1000 e100 mii floppy ext3 jbd
Mar 13 15:25:56 cfs1 kernel: CPU:    1
Mar 13 15:25:56 cfs1 kernel: EIP: 0060:[<e0c88da5>]
Not tainted VLI
Mar 13 15:25:56 cfs1 kernel: EFLAGS: 00010246   (2.6.20.1 #1)
Mar 13 15:25:56 cfs1 kernel: EIP is at gfs2_meta_indirect_buffer+0x4c/0x278 [gfs2] Mar 13 15:25:56 cfs1 kernel: eax: 00000000 ebx: 00012bf5 ecx: ce5a6dd4 edx: dc5eae00 Mar 13 15:25:56 cfs1 kernel: esi: 00000000 edi: 00000000 ebp: dc5ea9a8 esp: ce5a6d58
Mar 13 15:25:56 cfs1 kernel: ds: 007b   es: 007b   ss: 0068
Mar 13 15:25:56 cfs1 kernel: Process bonnie++ (pid: 5509,
ti=ce5a6000
task=dc320570 task.ti=ce5a6000)
Mar 13 15:25:56 cfs1 kernel: Stack: c156d274 ce5a6dd4 c016eb91 ce5a6dd4 00000000 dc5eae00 00000000 d6f34000 Mar 13 15:25:56 cfs1 kernel: 00000000 00000000 dc5ea9a8 d57794a8 ce5a6e08 e0c83ad3 00012bf5 00000000 Mar 13 15:25:56 cfs1 kernel: 00000000 ce5a6da0 00000000 00000000 dc5ea9a8 d57794a8 e0c84cc9 ce5a6dd4
Mar 13 15:25:56 cfs1 kernel: Call Trace:
Mar 13 15:25:56 cfs1 kernel:  [<c016eb91>] iget5_locked+0x3d/0x67
Mar 13 15:25:56 cfs1 kernel: [<e0c83ad3>] gfs2_inode_refresh+0x34/0xfe [gfs2] Mar 13 15:25:56 cfs1 kernel: [<e0c84cc9>]
gfs2_createi+0x12c/0x191 [gfs2]
Mar 13 15:25:56 cfs1 kernel: [<e0c8da3c>]
gfs2_create+0x5c/0x103 [gfs2]
Mar 13 15:25:56 cfs1 kernel: [<e0c84be7>]
gfs2_createi+0x4a/0x191 [gfs2]
Mar 13 15:25:56 cfs1 kernel: [<e0c820c4>]
gfs2_glock_nq_num+0x3f/0x64
[gfs2]
Mar 13 15:25:56 cfs1 kernel:  [<c016552e>] vfs_create+0xc3/0x126
Mar 13 15:25:56 cfs1 kernel: [<c01657f2>]
open_namei_create+0x47/0x88
Mar 13 15:25:56 cfs1 kernel:  [<c016597d>] open_namei+0x14a/0x539
Mar 13 15:25:56 cfs1 kernel:  [<c015d27b>] do_filp_open+0x25/0x39
Mar 13 15:25:56 cfs1 kernel: [<c01d200a>]
strncpy_from_user+0x3c/0x5b
Mar 13 15:25:56 cfs1 kernel:  [<c015d431>] get_unused_fd+0xa8/0xb1
Mar 13 15:25:56 cfs1 kernel:  [<c015d509>] do_sys_open+0x42/0xbe
Mar 13 15:25:56 cfs1 kernel:  [<c015d59f>] sys_open+0x1a/0x1c
Mar 13 15:25:56 cfs1 kernel:  [<c015d5dd>] sys_creat+0x1f/0x23
Mar 13 15:25:56 cfs1 kernel: [<c0103410>]
sysenter_past_esp+0x5d/0x81
Mar 13 15:25:56 cfs1 kernel:  =======================
Mar 13 15:25:56 cfs1 kernel: Code: 80 a8 01 00 00 89 44 24
1c 8b 85 f0
01 00 00 c7 44 24 20 00 00 00 00 89 54 24 14 85 c0 89 44 24
18 c7 44
24 10 00 00 00 00 75 04 <0f> 0b eb fe 83 7c 24 1c 00 75 04
0f 0b eb fe
8d 85 24 04 00 00
Mar 13 15:25:57 cfs1 kernel: EIP: [<e0c88da5>] gfs2_meta_indirect_buffer+0x4c/0x278 [gfs2] SS:ESP 0068:ce5a6d58 (2) Mar 13 17:00:30 cfs1 kernel: Call Trace: Mar 13 17:00:30 cfs1 kernel: [<e0badde1>]
gfs2_unlink+0x53/0xe0 [gfs2]
Mar 13 17:00:30 cfs1 kernel: [<e0baddc8>]
gfs2_unlink+0x3a/0xe0 [gfs2]
Mar 13 17:00:30 cfs1 kernel: [<e0badde1>]
gfs2_unlink+0x53/0xe0 [gfs2]
Mar 13 17:00:30 cfs1 kernel:  [<c0166572>] vfs_unlink+0xa1/0xc5
Mar 13 17:00:30 cfs1 kernel:  [<c016662b>] do_unlinkat+0x95/0xf5
Mar 13 17:00:30 cfs1 kernel:  [<c01187de>] scheduler_tick+0x8f/0x95
Mar 13 17:00:30 cfs1 kernel: [<c0103410>]
sysenter_past_esp+0x5d/0x81
Mar 13 17:00:30 cfs1 kernel:  =======================

-------------------------------------------------------------
-----------
--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster


--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux