----- Original Message ----- | Hello Folks, | RE: gfs:gfs_assert_i+0x67/0x92 seen when node joining cluster | | Has this been seen at other sites? | | Call Trace: | [<ffffffff888b2ffb>] :gfs:gfs_assert_i+0x67/0x92 | [<ffffffff888a0ed4>] :gfs:unlinked_scan_elements+0x99/0x180 | [<ffffffff88887a5f>] :gfs:gfs_dreread+0x87/0xc6 | [<ffffffff888acdc8>] :gfs:foreach_descriptor+0x229/0x305 | [<ffffffff888a6d02>] :gfs:fill_super+0x0/0x642 | [<ffffffff888ad12b>] :gfs:gfs_recover_dump+0xdd/0x14e | [<ffffffff888b1293>] :gfs:gfs_make_fs_rw+0xc0/0x11a | [<ffffffff888a6766>] :gfs:init_journal+0x279/0x34c | [<ffffffff888a7190>] :gfs:fill_super+0x48e/0x642 | [<ffffffff800e7461>] get_sb_bdev+0x10a/0x16c | [<ffffffff800e6dfe>] vfs_kern_mount+0x93/0x11a | [<ffffffff800e6ec7>] do_kern_mount+0x36/0x4d | [<ffffffff800f18c5>] do_mount+0x6a9/0x719 | [<ffffffff8008e202>] enqueue_task+0x41/0x56 | [<ffffffff80045ad3>] do_sock_read+0xcf/0x110 | [<ffffffff8022c620>] sock_aio_read+0x4f/0x5e | [<ffffffff8000cfdf>] do_sync_read+0xc7/0x104 | [<ffffffff800ceeb4>] zone_statistics+0x3e/0x6d | [<ffffffff8000f470>] __alloc_pages+0x78/0x308 | [<ffffffff8004c0df>] sys_mount+0x8a/0xcd | | Sep 18 04:09:51 hpium2 syslogd 1.4.1: restart. | Sep 18 04:09:51 hpium2 kernel: klogd 1.4.1, log source = /proc/kmsg started. | | Regards, | James Hofmeister Hewlett Packard Linux Engineering Resolution Team | Hi James, This might possibly be caused by corruption in the GFS file system. It's hard to say without detailed analysis of the GFS metadata. There was likely an "assert" message that appeared earlier; do you have that message? Regards, Bob Peterson Red Hat File Systems -- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster