RE: Kernel oops when GFS2 used as localfs

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Thanks Kevin! That helps.
 
- Sridhar


From: linux-cluster-bounces@xxxxxxxxxx [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Kevin Anderson
Sent: Saturday, March 17, 2007 2:37 AM
To: linux clustering
Subject: RE: Kernel oops when GFS2 used as localfs

Just opened up bugzilla 229831, you should be able to see it now.

Kevin

On Sat, 2007-03-17 at 00:11 -0600, David wrote:
Once you get it, feel free to share with all of us.

David 

-----Original Message-----
From: linux-cluster-bounces@xxxxxxxxxx
[mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Sridhar Ramaswamy
(srramasw)
Sent: Wednesday, March 14, 2007 12:20 PM
To: linux clustering
Subject: RE:  Kernel oops when GFS2 used as localfs


Hi folks,

This bugzilla 229831 seems "restricted" :(  Can someone please forward
its details? I'm interested in understanding the actual problem and its
proposed fix.

Meanwhile I'll try an upstream kernel. Should 2.6.21.rc1 be fine?

Thanks,
Sridhar

> -----Original Message-----
> From: linux-cluster-bounces@xxxxxxxxxx
> [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Wendy Cheng
> Sent: Wednesday, March 14, 2007 6:33 AM
> To: linux clustering
> Subject: Re:  Kernel oops when GFS2 used as localfs
> 
> Steven Whitehouse wrote:
> > Hi,
> >
> > This looks like Red Hat bugzilla 229831 and if so then
> there is already
> > a fix in Linus' upstream kernel and in the latest kernel build for 
> > Fedora (both 5 and 6),
> >
> >   
> yeah... agree. I missed the unlink part of the trace in previous post
> and thought it might have something to do with my ino number 
> hacking in 
> the lookup code.
> 
> -- Wendy
> > Steve.
> >
> > On Wed, 2007-03-14 at 00:15 -0700, Sridhar Ramaswamy
> (srramasw) wrote:
> >   
> >> Sure Wendy. Here it is,
> >>
> >> "fs/gfs2/inode.c" 1256
> >> struct inode *gfs2_ilookup(struct super_block *sb, struct
> gfs2_inum_host
> >> *inum)
> >> {
> >>         return ilookup5(sb, (unsigned long)inum->no_formal_ino,
> >>                         iget_test, inum);
> >> }
> >>
> >> static struct inode *gfs2_iget(struct super_block *sb, struct 
> >> gfs2_inum_host *inum) {
> >>         return iget5_locked(sb, (unsigned long)inum->no_formal_ino,
> >>                      iget_test, iget_set, inum);
> >> }
> >>
> >> BTW - this code is from kernel version 2.6.20.1.
> >>
> >> Thanks,
> >> Sridhar
> >>
> >>     
> >>> -----Original Message-----
> >>> From: linux-cluster-bounces@xxxxxxxxxx
> >>> [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Wendy Cheng
> >>> Sent: Tuesday, March 13, 2007 10:29 PM
> >>> To: linux clustering
> >>> Subject: Re:  Kernel oops when GFS2 used as localfs
> >>>
> >>> Sridhar Ramaswamy (srramasw) wrote:
> >>>
> >>>       
> >>>> Mar 13 15:25:56 cfs1 kernel: ------------[ cut here ]------------

> >>>> Mar 13 15:25:56 cfs1 kernel: kernel BUG at fs/gfs2/meta_io.c:474!
> >>>>         
> >>> I don't have time to pull kernel.org source code right
> now. Could you
> >>> cut-and-paste the following two routines (from fs/gfs2/inode.c):
> >>> gfs2_ilookup() and gfs2_iget() so we can look into this
> >>> quickly ? They 
> >>> are all one-liner so shouldn't be too much troubles. GFS2 
> >>> currently has 
> >>> ino issue with its lookup code.
> >>>
> >>> -- Wendy
> >>>
> >>>       
> >>>> Mar 13 15:25:56 cfs1 kernel: invalid opcode: 0000 [#1] Mar 13 
> >>>> 15:25:56 cfs1 kernel: SMP Mar 13 15:25:56 cfs1 kernel: Modules 
> >>>> linked in: lock_nolock gfs2 reiserfs nfsd exportfs nfs lockd 
> >>>> nfs_acl ipv6 parport_pc
> lp parport
> >>>> autofs4 sunrpc dm_mirror dm_mod button battery ac
> uhci_hcd ehci_hcd
> >>>> intel_rng rng_core i2c_i801 i2c_core e1000 e100 mii
> floppy ext3 jbd
> >>>> Mar 13 15:25:56 cfs1 kernel: CPU:    1
> >>>> Mar 13 15:25:56 cfs1 kernel: EIP:    0060:[<e0c88da5>]    
> >>>>         
> >>> Not tainted VLI
> >>>       
> >>>> Mar 13 15:25:56 cfs1 kernel: EFLAGS: 00010246   (2.6.20.1 #1)
> >>>> Mar 13 15:25:56 cfs1 kernel: EIP is at
> >>>> gfs2_meta_indirect_buffer+0x4c/0x278 [gfs2]
> >>>> Mar 13 15:25:56 cfs1 kernel: eax: 00000000   ebx: 
> 00012bf5   ecx: 
> >>>> ce5a6dd4   edx: dc5eae00
> >>>> Mar 13 15:25:56 cfs1 kernel: esi: 00000000   edi: 
> 00000000   ebp: 
> >>>> dc5ea9a8   esp: ce5a6d58
> >>>> Mar 13 15:25:56 cfs1 kernel: ds: 007b   es: 007b   ss: 0068
> >>>> Mar 13 15:25:56 cfs1 kernel: Process bonnie++ (pid: 5509,
> >>>>         
> >>> ti=ce5a6000
> >>>       
> >>>> task=dc320570 task.ti=ce5a6000)
> >>>> Mar 13 15:25:56 cfs1 kernel: Stack: c156d274 ce5a6dd4 c016eb91
> >>>> ce5a6dd4 00000000 dc5eae00 00000000 d6f34000
> >>>> Mar 13 15:25:56 cfs1 kernel:        00000000 00000000 dc5ea9a8 
> >>>> d57794a8 ce5a6e08 e0c83ad3 00012bf5 00000000
> >>>> Mar 13 15:25:56 cfs1 kernel:        00000000 ce5a6da0 00000000 
> >>>> 00000000 dc5ea9a8 d57794a8 e0c84cc9 ce5a6dd4
> >>>> Mar 13 15:25:56 cfs1 kernel: Call Trace:
> >>>> Mar 13 15:25:56 cfs1 kernel:  [<c016eb91>] iget5_locked+0x3d/0x67
> >>>> Mar 13 15:25:56 cfs1 kernel:  [<e0c83ad3>] 
> >>>> gfs2_inode_refresh+0x34/0xfe [gfs2]
> >>>> Mar 13 15:25:56 cfs1 kernel:  [<e0c84cc9>] 
> >>>>         
> >>> gfs2_createi+0x12c/0x191 [gfs2]
> >>>       
> >>>> Mar 13 15:25:56 cfs1 kernel:  [<e0c8da3c>]
> >>>>         
> >>> gfs2_create+0x5c/0x103 [gfs2]
> >>>       
> >>>> Mar 13 15:25:56 cfs1 kernel:  [<e0c84be7>]
> >>>>         
> >>> gfs2_createi+0x4a/0x191 [gfs2]
> >>>       
> >>>> Mar 13 15:25:56 cfs1 kernel:  [<e0c820c4>]
> >>>>         
> >>> gfs2_glock_nq_num+0x3f/0x64
> >>>       
> >>>> [gfs2]
> >>>> Mar 13 15:25:56 cfs1 kernel:  [<c016552e>] vfs_create+0xc3/0x126 
> >>>> Mar 13 15:25:56 cfs1 kernel:  [<c01657f2>]
> >>>>         
> >>> open_namei_create+0x47/0x88
> >>>       
> >>>> Mar 13 15:25:56 cfs1 kernel:  [<c016597d>] open_namei+0x14a/0x539

> >>>> Mar 13 15:25:56 cfs1 kernel:  [<c015d27b>] do_filp_open+0x25/0x39

> >>>> Mar 13 15:25:56 cfs1 kernel:  [<c01d200a>]
> >>>>         
> >>> strncpy_from_user+0x3c/0x5b
> >>>       
> >>>> Mar 13 15:25:56 cfs1 kernel:  [<c015d431>]
> get_unused_fd+0xa8/0xb1
> >>>> Mar 13 15:25:56 cfs1 kernel:  [<c015d509>] do_sys_open+0x42/0xbe 
> >>>> Mar 13 15:25:56 cfs1 kernel:  [<c015d59f>] sys_open+0x1a/0x1c Mar

> >>>> 13 15:25:56 cfs1 kernel:  [<c015d5dd>] sys_creat+0x1f/0x23 Mar 13

> >>>> 15:25:56 cfs1 kernel:  [<c0103410>]
> >>>>         
> >>> sysenter_past_esp+0x5d/0x81
> >>>       
> >>>> Mar 13 15:25:56 cfs1 kernel:  ======================= Mar 13 
> >>>> 15:25:56 cfs1 kernel: Code: 80 a8 01 00 00 89 44 24
> >>>>         
> >>> 1c 8b 85 f0
> >>>       
> >>>> 01 00 00 c7 44 24 20 00 00 00 00 89 54 24 14 85 c0 89 44 24
> >>>>         
> >>> 18 c7 44
> >>>       
> >>>> 24 10 00 00 00 00 75 04 <0f> 0b eb fe 83 7c 24 1c 00 75 04
> >>>>         
> >>> 0f 0b eb fe
> >>>       
> >>>> 8d 85 24 04 00 00
> >>>> Mar 13 15:25:57 cfs1 kernel: EIP: [<e0c88da5>]
> >>>> gfs2_meta_indirect_buffer+0x4c/0x278 [gfs2] SS:ESP 0068:ce5a6d58
> >>>>  
> >>>>  
> >>>> (2)
> >>>>  
> >>>> Mar 13 17:00:30 cfs1 kernel: Call Trace:
> >>>> Mar 13 17:00:30 cfs1 kernel:  [<e0badde1>]
> >>>>         
> >>> gfs2_unlink+0x53/0xe0 [gfs2]
> >>>       
> >>>> Mar 13 17:00:30 cfs1 kernel:  [<e0baddc8>]
> >>>>         
> >>> gfs2_unlink+0x3a/0xe0 [gfs2]
> >>>       
> >>>> Mar 13 17:00:30 cfs1 kernel:  [<e0badde1>]
> >>>>         
> >>> gfs2_unlink+0x53/0xe0 [gfs2]
> >>>       
> >>>> Mar 13 17:00:30 cfs1 kernel:  [<c0166572>] vfs_unlink+0xa1/0xc5 
> >>>> Mar 13 17:00:30 cfs1 kernel:  [<c016662b>] do_unlinkat+0x95/0xf5 
> >>>> Mar 13 17:00:30 cfs1 kernel:  [<c01187de>]
> scheduler_tick+0x8f/0x95
> >>>> Mar 13 17:00:30 cfs1 kernel:  [<c0103410>]
> >>>>         
> >>> sysenter_past_esp+0x5d/0x81
> >>>       
> >>>> Mar 13 17:00:30 cfs1 kernel:  =======================
> >>>>
> >>>> -------------------------------------------------------------
> >>>>         
> >>> -----------
> >>>       
> >>>> --
> >>>> Linux-cluster mailing list
> >>>> Linux-cluster@xxxxxxxxxx 
> >>>> https://www.redhat.com/mailman/listinfo/linux-cluster
> >>>>
> >>>>         
> >>> --
> >>> Linux-cluster mailing list
> >>> Linux-cluster@xxxxxxxxxx 
> >>> https://www.redhat.com/mailman/listinfo/linux-cluster
> >>>
> >>>       
> >> --
> >> Linux-cluster mailing list
> >> Linux-cluster@xxxxxxxxxx 
> >> https://www.redhat.com/mailman/listinfo/linux-cluster
> >>     
> >
> >   
> 
> --
> Linux-cluster mailing list
> Linux-cluster@xxxxxxxxxx 
> https://www.redhat.com/mailman/listinfo/linux-cluster
> 

--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster
--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux