Once you get it, feel free to share with all of us. David -----Original Message----- From: linux-cluster-bounces@xxxxxxxxxx [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Sridhar Ramaswamy (srramasw) Sent: Wednesday, March 14, 2007 12:20 PM To: linux clustering Subject: RE: Kernel oops when GFS2 used as localfs Hi folks, This bugzilla 229831 seems "restricted" :( Can someone please forward its details? I'm interested in understanding the actual problem and its proposed fix. Meanwhile I'll try an upstream kernel. Should 2.6.21.rc1 be fine? Thanks, Sridhar > -----Original Message----- > From: linux-cluster-bounces@xxxxxxxxxx > [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Wendy Cheng > Sent: Wednesday, March 14, 2007 6:33 AM > To: linux clustering > Subject: Re: Kernel oops when GFS2 used as localfs > > Steven Whitehouse wrote: > > Hi, > > > > This looks like Red Hat bugzilla 229831 and if so then > there is already > > a fix in Linus' upstream kernel and in the latest kernel build for > > Fedora (both 5 and 6), > > > > > yeah... agree. I missed the unlink part of the trace in previous post > and thought it might have something to do with my ino number > hacking in > the lookup code. > > -- Wendy > > Steve. > > > > On Wed, 2007-03-14 at 00:15 -0700, Sridhar Ramaswamy > (srramasw) wrote: > > > >> Sure Wendy. Here it is, > >> > >> "fs/gfs2/inode.c" 1256 > >> struct inode *gfs2_ilookup(struct super_block *sb, struct > gfs2_inum_host > >> *inum) > >> { > >> return ilookup5(sb, (unsigned long)inum->no_formal_ino, > >> iget_test, inum); > >> } > >> > >> static struct inode *gfs2_iget(struct super_block *sb, struct > >> gfs2_inum_host *inum) { > >> return iget5_locked(sb, (unsigned long)inum->no_formal_ino, > >> iget_test, iget_set, inum); > >> } > >> > >> BTW - this code is from kernel version 2.6.20.1. > >> > >> Thanks, > >> Sridhar > >> > >> > >>> -----Original Message----- > >>> From: linux-cluster-bounces@xxxxxxxxxx > >>> [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Wendy Cheng > >>> Sent: Tuesday, March 13, 2007 10:29 PM > >>> To: linux clustering > >>> Subject: Re: Kernel oops when GFS2 used as localfs > >>> > >>> Sridhar Ramaswamy (srramasw) wrote: > >>> > >>> > >>>> Mar 13 15:25:56 cfs1 kernel: ------------[ cut here ]------------ > >>>> Mar 13 15:25:56 cfs1 kernel: kernel BUG at fs/gfs2/meta_io.c:474! > >>>> > >>> I don't have time to pull kernel.org source code right > now. Could you > >>> cut-and-paste the following two routines (from fs/gfs2/inode.c): > >>> gfs2_ilookup() and gfs2_iget() so we can look into this > >>> quickly ? They > >>> are all one-liner so shouldn't be too much troubles. GFS2 > >>> currently has > >>> ino issue with its lookup code. > >>> > >>> -- Wendy > >>> > >>> > >>>> Mar 13 15:25:56 cfs1 kernel: invalid opcode: 0000 [#1] Mar 13 > >>>> 15:25:56 cfs1 kernel: SMP Mar 13 15:25:56 cfs1 kernel: Modules > >>>> linked in: lock_nolock gfs2 reiserfs nfsd exportfs nfs lockd > >>>> nfs_acl ipv6 parport_pc > lp parport > >>>> autofs4 sunrpc dm_mirror dm_mod button battery ac > uhci_hcd ehci_hcd > >>>> intel_rng rng_core i2c_i801 i2c_core e1000 e100 mii > floppy ext3 jbd > >>>> Mar 13 15:25:56 cfs1 kernel: CPU: 1 > >>>> Mar 13 15:25:56 cfs1 kernel: EIP: 0060:[<e0c88da5>] > >>>> > >>> Not tainted VLI > >>> > >>>> Mar 13 15:25:56 cfs1 kernel: EFLAGS: 00010246 (2.6.20.1 #1) > >>>> Mar 13 15:25:56 cfs1 kernel: EIP is at > >>>> gfs2_meta_indirect_buffer+0x4c/0x278 [gfs2] > >>>> Mar 13 15:25:56 cfs1 kernel: eax: 00000000 ebx: > 00012bf5 ecx: > >>>> ce5a6dd4 edx: dc5eae00 > >>>> Mar 13 15:25:56 cfs1 kernel: esi: 00000000 edi: > 00000000 ebp: > >>>> dc5ea9a8 esp: ce5a6d58 > >>>> Mar 13 15:25:56 cfs1 kernel: ds: 007b es: 007b ss: 0068 > >>>> Mar 13 15:25:56 cfs1 kernel: Process bonnie++ (pid: 5509, > >>>> > >>> ti=ce5a6000 > >>> > >>>> task=dc320570 task.ti=ce5a6000) > >>>> Mar 13 15:25:56 cfs1 kernel: Stack: c156d274 ce5a6dd4 c016eb91 > >>>> ce5a6dd4 00000000 dc5eae00 00000000 d6f34000 > >>>> Mar 13 15:25:56 cfs1 kernel: 00000000 00000000 dc5ea9a8 > >>>> d57794a8 ce5a6e08 e0c83ad3 00012bf5 00000000 > >>>> Mar 13 15:25:56 cfs1 kernel: 00000000 ce5a6da0 00000000 > >>>> 00000000 dc5ea9a8 d57794a8 e0c84cc9 ce5a6dd4 > >>>> Mar 13 15:25:56 cfs1 kernel: Call Trace: > >>>> Mar 13 15:25:56 cfs1 kernel: [<c016eb91>] iget5_locked+0x3d/0x67 > >>>> Mar 13 15:25:56 cfs1 kernel: [<e0c83ad3>] > >>>> gfs2_inode_refresh+0x34/0xfe [gfs2] > >>>> Mar 13 15:25:56 cfs1 kernel: [<e0c84cc9>] > >>>> > >>> gfs2_createi+0x12c/0x191 [gfs2] > >>> > >>>> Mar 13 15:25:56 cfs1 kernel: [<e0c8da3c>] > >>>> > >>> gfs2_create+0x5c/0x103 [gfs2] > >>> > >>>> Mar 13 15:25:56 cfs1 kernel: [<e0c84be7>] > >>>> > >>> gfs2_createi+0x4a/0x191 [gfs2] > >>> > >>>> Mar 13 15:25:56 cfs1 kernel: [<e0c820c4>] > >>>> > >>> gfs2_glock_nq_num+0x3f/0x64 > >>> > >>>> [gfs2] > >>>> Mar 13 15:25:56 cfs1 kernel: [<c016552e>] vfs_create+0xc3/0x126 > >>>> Mar 13 15:25:56 cfs1 kernel: [<c01657f2>] > >>>> > >>> open_namei_create+0x47/0x88 > >>> > >>>> Mar 13 15:25:56 cfs1 kernel: [<c016597d>] open_namei+0x14a/0x539 > >>>> Mar 13 15:25:56 cfs1 kernel: [<c015d27b>] do_filp_open+0x25/0x39 > >>>> Mar 13 15:25:56 cfs1 kernel: [<c01d200a>] > >>>> > >>> strncpy_from_user+0x3c/0x5b > >>> > >>>> Mar 13 15:25:56 cfs1 kernel: [<c015d431>] > get_unused_fd+0xa8/0xb1 > >>>> Mar 13 15:25:56 cfs1 kernel: [<c015d509>] do_sys_open+0x42/0xbe > >>>> Mar 13 15:25:56 cfs1 kernel: [<c015d59f>] sys_open+0x1a/0x1c Mar > >>>> 13 15:25:56 cfs1 kernel: [<c015d5dd>] sys_creat+0x1f/0x23 Mar 13 > >>>> 15:25:56 cfs1 kernel: [<c0103410>] > >>>> > >>> sysenter_past_esp+0x5d/0x81 > >>> > >>>> Mar 13 15:25:56 cfs1 kernel: ======================= Mar 13 > >>>> 15:25:56 cfs1 kernel: Code: 80 a8 01 00 00 89 44 24 > >>>> > >>> 1c 8b 85 f0 > >>> > >>>> 01 00 00 c7 44 24 20 00 00 00 00 89 54 24 14 85 c0 89 44 24 > >>>> > >>> 18 c7 44 > >>> > >>>> 24 10 00 00 00 00 75 04 <0f> 0b eb fe 83 7c 24 1c 00 75 04 > >>>> > >>> 0f 0b eb fe > >>> > >>>> 8d 85 24 04 00 00 > >>>> Mar 13 15:25:57 cfs1 kernel: EIP: [<e0c88da5>] > >>>> gfs2_meta_indirect_buffer+0x4c/0x278 [gfs2] SS:ESP 0068:ce5a6d58 > >>>> > >>>> > >>>> (2) > >>>> > >>>> Mar 13 17:00:30 cfs1 kernel: Call Trace: > >>>> Mar 13 17:00:30 cfs1 kernel: [<e0badde1>] > >>>> > >>> gfs2_unlink+0x53/0xe0 [gfs2] > >>> > >>>> Mar 13 17:00:30 cfs1 kernel: [<e0baddc8>] > >>>> > >>> gfs2_unlink+0x3a/0xe0 [gfs2] > >>> > >>>> Mar 13 17:00:30 cfs1 kernel: [<e0badde1>] > >>>> > >>> gfs2_unlink+0x53/0xe0 [gfs2] > >>> > >>>> Mar 13 17:00:30 cfs1 kernel: [<c0166572>] vfs_unlink+0xa1/0xc5 > >>>> Mar 13 17:00:30 cfs1 kernel: [<c016662b>] do_unlinkat+0x95/0xf5 > >>>> Mar 13 17:00:30 cfs1 kernel: [<c01187de>] > scheduler_tick+0x8f/0x95 > >>>> Mar 13 17:00:30 cfs1 kernel: [<c0103410>] > >>>> > >>> sysenter_past_esp+0x5d/0x81 > >>> > >>>> Mar 13 17:00:30 cfs1 kernel: ======================= > >>>> > >>>> ------------------------------------------------------------- > >>>> > >>> ----------- > >>> > >>>> -- > >>>> Linux-cluster mailing list > >>>> Linux-cluster@xxxxxxxxxx > >>>> https://www.redhat.com/mailman/listinfo/linux-cluster > >>>> > >>>> > >>> -- > >>> Linux-cluster mailing list > >>> Linux-cluster@xxxxxxxxxx > >>> https://www.redhat.com/mailman/listinfo/linux-cluster > >>> > >>> > >> -- > >> Linux-cluster mailing list > >> Linux-cluster@xxxxxxxxxx > >> https://www.redhat.com/mailman/listinfo/linux-cluster > >> > > > > > > -- > Linux-cluster mailing list > Linux-cluster@xxxxxxxxxx > https://www.redhat.com/mailman/listinfo/linux-cluster > -- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster -- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster