Cannot sync RAID5 anymore - MCE problem?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello everybody.

I have a strange and severe problem with my Raid-Array and I have to
contact "experts" before I can continue with a clear conscience.
I wanted to exchange 2 disks in my Raid5 array by two newer ones. So I
connected the new to my machine partitioned the correct layout added one
of them as a Spare and "--faild" the partition of the disk I wanted to
remove first. Rebulding of the array started immediately and finished fine.
Now I took the two old disks out and put the new ones in. By removing the
other disk from my array I degraded it. After booting I added the correct
partition of my new drive to the Raid5 and waited for the syncing to
finish ...
but it didn't. Is crashed my whole machine with an MCE:
CPU 0: Machine check exception: 4 bank 4 b20000000000070f0f
...
Kernel Panic - not synching: machine check

The reason why I write an MCE problem to the Software-Raid list is that
this problem is very reproducible and always happens when resyncing of my
array has finished 24.9%. I tried it about ten times so I am really sure
that there is some connection to resyncing since this problem does not
seem to appear under different conditions anymore.
I tried to do an rsync-backup of my raid-array which lead the the same
crash. After that I observed that this crash has occured when copying a
not so important Backup of something else. I deleted that old Backup and
since that my problem seems to ONLY occur if I try to resync my array.

I am running Gentoo on an AMD64 3200+ and K8N Neo4 Platinum and my problem
seems to be similar to the problems of these guys:
http://kerneltrap.org/node/4993
but somehow related to resyncing. I have reiserfs on my array and
successfully completed a "reiserfsck --rebuild-tree".
I think it is not important but it might be good to mention that I use
LVM, too.

I have also tried to resync the array to my old disk (with the second new
one removed), but that leads to the same problem.

I have tried several things like removing one RAM module or using
different RAM-Banks I checked for leaking caps I tried without DMA, tried
different kernels and played with some kernel options.

Is there a way to figure out what hardware seems to be the problem?
My hardware worked flawlessly for over 1.5 years if I did not break
something while physically mounting the disks or cleaning dust out of the
case it can only be a problem of the first new harddrive (which is
unfortunately part of my degraded raid-array already). Is it possible that
an SATA1 Cable on a SATA2 capable controller connected to a SATA2 capable
disk leads to such errors?

Since I was able to copy my data I think it is in perfect condition, but
there seems to be a problem on the array in the "empty"-part. Does anybody
know a way how to over- or rewrite the empty blocks of a
reiserfs-partition? Or some tool to find/correct disk-problems. (I tried
reiserfsck but that does not find anything)

What is the smartest way for my to proceed to get my degraded array
redundant again?
I could delete the whole array, try set it up identically again and recopy
the data, but if it is really a hardware problem that would be a waste of
time.

Thanks in advance ...
	Ramin
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux