Sure. I don't have any breakthru' yet. Also I didn't got any further details on this bugzilla contents. - Sridhar > -----Original Message----- > From: linux-cluster-bounces@xxxxxxxxxx > [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of David > Sent: Friday, March 16, 2007 11:11 PM > To: 'linux clustering' > Subject: RE: Kernel oops when GFS2 used as localfs > > Once you get it, feel free to share with all of us. > > David > > -----Original Message----- > From: linux-cluster-bounces@xxxxxxxxxx > [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of > Sridhar Ramaswamy > (srramasw) > Sent: Wednesday, March 14, 2007 12:20 PM > To: linux clustering > Subject: RE: Kernel oops when GFS2 used as localfs > > > Hi folks, > > This bugzilla 229831 seems "restricted" :( Can someone please forward > its details? I'm interested in understanding the actual > problem and its > proposed fix. > > Meanwhile I'll try an upstream kernel. Should 2.6.21.rc1 be fine? > > Thanks, > Sridhar > > > -----Original Message----- > > From: linux-cluster-bounces@xxxxxxxxxx > > [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Wendy Cheng > > Sent: Wednesday, March 14, 2007 6:33 AM > > To: linux clustering > > Subject: Re: Kernel oops when GFS2 used as localfs > > > > Steven Whitehouse wrote: > > > Hi, > > > > > > This looks like Red Hat bugzilla 229831 and if so then > > there is already > > > a fix in Linus' upstream kernel and in the latest kernel > build for > > > Fedora (both 5 and 6), > > > > > > > > yeah... agree. I missed the unlink part of the trace in > previous post > > and thought it might have something to do with my ino number > > hacking in > > the lookup code. > > > > -- Wendy > > > Steve. > > > > > > On Wed, 2007-03-14 at 00:15 -0700, Sridhar Ramaswamy > > (srramasw) wrote: > > > > > >> Sure Wendy. Here it is, > > >> > > >> "fs/gfs2/inode.c" 1256 > > >> struct inode *gfs2_ilookup(struct super_block *sb, struct > > gfs2_inum_host > > >> *inum) > > >> { > > >> return ilookup5(sb, (unsigned long)inum->no_formal_ino, > > >> iget_test, inum); > > >> } > > >> > > >> static struct inode *gfs2_iget(struct super_block *sb, struct > > >> gfs2_inum_host *inum) { > > >> return iget5_locked(sb, (unsigned > long)inum->no_formal_ino, > > >> iget_test, iget_set, inum); > > >> } > > >> > > >> BTW - this code is from kernel version 2.6.20.1. > > >> > > >> Thanks, > > >> Sridhar > > >> > > >> > > >>> -----Original Message----- > > >>> From: linux-cluster-bounces@xxxxxxxxxx > > >>> [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of > Wendy Cheng > > >>> Sent: Tuesday, March 13, 2007 10:29 PM > > >>> To: linux clustering > > >>> Subject: Re: Kernel oops when GFS2 used > as localfs > > >>> > > >>> Sridhar Ramaswamy (srramasw) wrote: > > >>> > > >>> > > >>>> Mar 13 15:25:56 cfs1 kernel: ------------[ cut here > ]------------ > > > >>>> Mar 13 15:25:56 cfs1 kernel: kernel BUG at > fs/gfs2/meta_io.c:474! > > >>>> > > >>> I don't have time to pull kernel.org source code right > > now. Could you > > >>> cut-and-paste the following two routines (from fs/gfs2/inode.c): > > >>> gfs2_ilookup() and gfs2_iget() so we can look into this > > >>> quickly ? They > > >>> are all one-liner so shouldn't be too much troubles. GFS2 > > >>> currently has > > >>> ino issue with its lookup code. > > >>> > > >>> -- Wendy > > >>> > > >>> > > >>>> Mar 13 15:25:56 cfs1 kernel: invalid opcode: 0000 [#1] Mar 13 > > >>>> 15:25:56 cfs1 kernel: SMP Mar 13 15:25:56 cfs1 kernel: Modules > > >>>> linked in: lock_nolock gfs2 reiserfs nfsd exportfs nfs lockd > > >>>> nfs_acl ipv6 parport_pc > > lp parport > > >>>> autofs4 sunrpc dm_mirror dm_mod button battery ac > > uhci_hcd ehci_hcd > > >>>> intel_rng rng_core i2c_i801 i2c_core e1000 e100 mii > > floppy ext3 jbd > > >>>> Mar 13 15:25:56 cfs1 kernel: CPU: 1 > > >>>> Mar 13 15:25:56 cfs1 kernel: EIP: 0060:[<e0c88da5>] > > >>>> > > >>> Not tainted VLI > > >>> > > >>>> Mar 13 15:25:56 cfs1 kernel: EFLAGS: 00010246 (2.6.20.1 #1) > > >>>> Mar 13 15:25:56 cfs1 kernel: EIP is at > > >>>> gfs2_meta_indirect_buffer+0x4c/0x278 [gfs2] > > >>>> Mar 13 15:25:56 cfs1 kernel: eax: 00000000 ebx: > > 00012bf5 ecx: > > >>>> ce5a6dd4 edx: dc5eae00 > > >>>> Mar 13 15:25:56 cfs1 kernel: esi: 00000000 edi: > > 00000000 ebp: > > >>>> dc5ea9a8 esp: ce5a6d58 > > >>>> Mar 13 15:25:56 cfs1 kernel: ds: 007b es: 007b ss: 0068 > > >>>> Mar 13 15:25:56 cfs1 kernel: Process bonnie++ (pid: 5509, > > >>>> > > >>> ti=ce5a6000 > > >>> > > >>>> task=dc320570 task.ti=ce5a6000) > > >>>> Mar 13 15:25:56 cfs1 kernel: Stack: c156d274 ce5a6dd4 c016eb91 > > >>>> ce5a6dd4 00000000 dc5eae00 00000000 d6f34000 > > >>>> Mar 13 15:25:56 cfs1 kernel: 00000000 00000000 dc5ea9a8 > > >>>> d57794a8 ce5a6e08 e0c83ad3 00012bf5 00000000 > > >>>> Mar 13 15:25:56 cfs1 kernel: 00000000 ce5a6da0 00000000 > > >>>> 00000000 dc5ea9a8 d57794a8 e0c84cc9 ce5a6dd4 > > >>>> Mar 13 15:25:56 cfs1 kernel: Call Trace: > > >>>> Mar 13 15:25:56 cfs1 kernel: [<c016eb91>] > iget5_locked+0x3d/0x67 > > >>>> Mar 13 15:25:56 cfs1 kernel: [<e0c83ad3>] > > >>>> gfs2_inode_refresh+0x34/0xfe [gfs2] > > >>>> Mar 13 15:25:56 cfs1 kernel: [<e0c84cc9>] > > >>>> > > >>> gfs2_createi+0x12c/0x191 [gfs2] > > >>> > > >>>> Mar 13 15:25:56 cfs1 kernel: [<e0c8da3c>] > > >>>> > > >>> gfs2_create+0x5c/0x103 [gfs2] > > >>> > > >>>> Mar 13 15:25:56 cfs1 kernel: [<e0c84be7>] > > >>>> > > >>> gfs2_createi+0x4a/0x191 [gfs2] > > >>> > > >>>> Mar 13 15:25:56 cfs1 kernel: [<e0c820c4>] > > >>>> > > >>> gfs2_glock_nq_num+0x3f/0x64 > > >>> > > >>>> [gfs2] > > >>>> Mar 13 15:25:56 cfs1 kernel: [<c016552e>] > vfs_create+0xc3/0x126 > > >>>> Mar 13 15:25:56 cfs1 kernel: [<c01657f2>] > > >>>> > > >>> open_namei_create+0x47/0x88 > > >>> > > >>>> Mar 13 15:25:56 cfs1 kernel: [<c016597d>] > open_namei+0x14a/0x539 > > > >>>> Mar 13 15:25:56 cfs1 kernel: [<c015d27b>] > do_filp_open+0x25/0x39 > > > >>>> Mar 13 15:25:56 cfs1 kernel: [<c01d200a>] > > >>>> > > >>> strncpy_from_user+0x3c/0x5b > > >>> > > >>>> Mar 13 15:25:56 cfs1 kernel: [<c015d431>] > > get_unused_fd+0xa8/0xb1 > > >>>> Mar 13 15:25:56 cfs1 kernel: [<c015d509>] > do_sys_open+0x42/0xbe > > >>>> Mar 13 15:25:56 cfs1 kernel: [<c015d59f>] > sys_open+0x1a/0x1c Mar > > > >>>> 13 15:25:56 cfs1 kernel: [<c015d5dd>] > sys_creat+0x1f/0x23 Mar 13 > > > >>>> 15:25:56 cfs1 kernel: [<c0103410>] > > >>>> > > >>> sysenter_past_esp+0x5d/0x81 > > >>> > > >>>> Mar 13 15:25:56 cfs1 kernel: ======================= Mar 13 > > >>>> 15:25:56 cfs1 kernel: Code: 80 a8 01 00 00 89 44 24 > > >>>> > > >>> 1c 8b 85 f0 > > >>> > > >>>> 01 00 00 c7 44 24 20 00 00 00 00 89 54 24 14 85 c0 89 44 24 > > >>>> > > >>> 18 c7 44 > > >>> > > >>>> 24 10 00 00 00 00 75 04 <0f> 0b eb fe 83 7c 24 1c 00 75 04 > > >>>> > > >>> 0f 0b eb fe > > >>> > > >>>> 8d 85 24 04 00 00 > > >>>> Mar 13 15:25:57 cfs1 kernel: EIP: [<e0c88da5>] > > >>>> gfs2_meta_indirect_buffer+0x4c/0x278 [gfs2] SS:ESP > 0068:ce5a6d58 > > >>>> > > >>>> > > >>>> (2) > > >>>> > > >>>> Mar 13 17:00:30 cfs1 kernel: Call Trace: > > >>>> Mar 13 17:00:30 cfs1 kernel: [<e0badde1>] > > >>>> > > >>> gfs2_unlink+0x53/0xe0 [gfs2] > > >>> > > >>>> Mar 13 17:00:30 cfs1 kernel: [<e0baddc8>] > > >>>> > > >>> gfs2_unlink+0x3a/0xe0 [gfs2] > > >>> > > >>>> Mar 13 17:00:30 cfs1 kernel: [<e0badde1>] > > >>>> > > >>> gfs2_unlink+0x53/0xe0 [gfs2] > > >>> > > >>>> Mar 13 17:00:30 cfs1 kernel: [<c0166572>] > vfs_unlink+0xa1/0xc5 > > >>>> Mar 13 17:00:30 cfs1 kernel: [<c016662b>] > do_unlinkat+0x95/0xf5 > > >>>> Mar 13 17:00:30 cfs1 kernel: [<c01187de>] > > scheduler_tick+0x8f/0x95 > > >>>> Mar 13 17:00:30 cfs1 kernel: [<c0103410>] > > >>>> > > >>> sysenter_past_esp+0x5d/0x81 > > >>> > > >>>> Mar 13 17:00:30 cfs1 kernel: ======================= > > >>>> > > >>>> ------------------------------------------------------------- > > >>>> > > >>> ----------- > > >>> > > >>>> -- > > >>>> Linux-cluster mailing list > > >>>> Linux-cluster@xxxxxxxxxx > > >>>> https://www.redhat.com/mailman/listinfo/linux-cluster > > >>>> > > >>>> > > >>> -- > > >>> Linux-cluster mailing list > > >>> Linux-cluster@xxxxxxxxxx > > >>> https://www.redhat.com/mailman/listinfo/linux-cluster > > >>> > > >>> > > >> -- > > >> Linux-cluster mailing list > > >> Linux-cluster@xxxxxxxxxx > > >> https://www.redhat.com/mailman/listinfo/linux-cluster > > >> > > > > > > > > > > -- > > Linux-cluster mailing list > > Linux-cluster@xxxxxxxxxx > > https://www.redhat.com/mailman/listinfo/linux-cluster > > > > -- > Linux-cluster mailing list > Linux-cluster@xxxxxxxxxx > https://www.redhat.com/mailman/listinfo/linux-cluster > > -- > Linux-cluster mailing list > Linux-cluster@xxxxxxxxxx > https://www.redhat.com/mailman/listinfo/linux-cluster > -- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster