On Friday 25 May 2007 06:55:00 David Chinner wrote: > Oh, did you look at your logs and find that XFS had spammed them > about writes that were failing? The first message after the incident: May 24 01:53:50 hq kernel: Filesystem "loop1": XFS internal error xfs_btree_check_sblock at line 336 of file fs/xfs/xfs_btree.c. Caller 0xf8ac14f8 May 24 01:53:50 hq kernel: <f8adae69> xfs_btree_check_sblock+0x4f/0xc2 [xfs] <f8ac14f8> xfs_alloc_lookup+0x34e/0x47b [xfs] May 24 01:53:50 HF kernel: <f8ac14f8> xfs_alloc_lookup+0x34e/0x47b [xfs] <f8b1a9c7> kmem_zone_zalloc+0x1b/0x43 [xfs] May 24 01:53:50 hq kernel: <f8abe645> xfs_alloc_ag_vextent+0x24d/0x1110 [xfs] <f8ac0647> xfs_alloc_vextent+0x3bd/0x53b [xfs] May 24 01:53:50 hq kernel: <f8ad2f7e> xfs_bmapi+0x1ac4/0x23cd [xfs] <f8acab97> xfs_bmap_search_multi_extents+0x8e/0xd8 [xfs] May 24 01:53:50 hq kernel: <f8b00001> xlog_dealloc_log+0x49/0xea [xfs] <f8afdaee> xfs_iomap_write_allocate+0x2d9/0x58b [xfs] May 24 01:53:50 hq kernel: <f8afc3ae> xfs_iomap+0x60e/0x82d [xfs] <c0113bc8> __wake_up_common+0x39/0x59 May 24 01:53:50 hq kernel: <f8b1ae11> xfs_map_blocks+0x39/0x6c [xfs] <f8b1bd7b> xfs_page_state_convert+0x644/0xf9c [xfs] May 24 01:53:50 hq kernel: <c036f384> schedule+0x5d1/0xf4d <f8b1c780> xfs_vm_writepage+0x0/0xe0 [xfs] May 24 01:53:50 hq kernel: <f8b1c7d7> xfs_vm_writepage+0x57/0xe0 [xfs] <c01830e8> mpage_writepages+0x1fb/0x3bb May 24 01:53:50 hq kernel: <c0183020> mpage_writepages+0x133/0x3bb <f8b1c780> xfs_vm_writepage+0x0/0xe0 [xfs] May 24 01:53:50 hq kernel: <c0147bb3> do_writepages+0x35/0x3b <c018135c> __writeback_single_inode+0x88/0x387 May 24 01:53:50 hq kernel: <c01819b7> sync_sb_inodes+0x1b4/0x2a8 <c0181c63> writeback_inodes+0x63/0xdc May 24 01:53:50 hq kernel: <c0147943> background_writeout+0x66/0x9f <c01482b3> pdflush+0x0/0x1ad May 24 01:53:50 hq kernel: <c01483a2> pdflush+0xef/0x1ad <c01478dd> background_writeout+0x0/0x9f May 24 01:53:50 hq kernel: <c012d10b> kthread+0xc2/0xc6 <c012d049> kthread+0x0/0xc6 May 24 01:53:50 hq kernel: <c0100dd5> kernel_thread_helper+0x5/0xb ..and I've spammed such messages. This "internal error" isn't a good reason to shut down the file system? I think if there's a sign of corrupted file system, the first thing we should do is to stop writes (or the entire FS) and let the admin to examine the situation. I'm not talking about my case where the md raid5 was a braindead, I'm talking about general situations. -- d - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html