Re: Re: Raid5 issues

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]



Toby Bluhm wrote:
Ruslan Sivak wrote:
Toby Bluhm wrote:
Ruslan Sivak wrote:
Feizhou wrote:
Ruslan Sivak wrote:
Feizhou wrote:

I do have a SIL3114 chipset, and I think it's supposed to be supported by device mapper. When I go to rescue mode, I see it loading the driver for SIL3112, but nothing appears under /dev/mapper except control. Are there instructions somewhere on getting it to use my controller's raid?

Your controller only has a bios chip. It has no raid processing capability at all.

You need to use mdadm. anaconda should be able to let you create to mirrors and then create a third array that stripes those md devices,
_______________________________________________
Anaconda doesn't let me create a stripe raid set on top of a mirror set. And it doesn't detect it when I do it manually. Also the bios chip presents additional issues. I believe when I don't have a raid array set up, it won't boot at all. When I have it on raid10, I had trouble booting, and when I have it on concatenation, everything works fine, until a drive is replaced. At that point, i have to recreate the array, as concatenation is not a fault tolerant set, and at this point I seem to lose all my data.

It won't boot at all without a raid array setup? That sounds really funny.

Actually I'm not 100% sure on this, but I think this is the case. I believe the first time I set it up as a raid10, assuming that linux will just ignore it. I installed centos by putting boot on a raid1, and root on LVM over 2 raid1 sets. I had trouble getting it to boot.
Is there a way to get it to use the raid that's part of the bios chip?

Repeat after me. There is no raid that is part of the bios chip. It is just a simple table.
Yes, I know this is fakeraid, aka softraid, but I was hoping that using the drivers would make it easier to support raid 10 then with mdadm, which seems to be impossible to get to work with the installer. I'm not even sure why the raid10 personality is not loaded, as it seems to have been part of the mdadm since version 1.7.
Something about device mapper?


You need the fake raid driver dmraid if you are going to set up stuff in the bios. What version of centos are you trying to install? libata in Centos 5 should support this without having to resort to the ide drivers.
_________________________________
I'm trying to install centos 5 - the latest. How would I go about using dmraid and/or libata? The installer picks up the drives as individual drives. There is a drive on the silicon image website, but it's for RHEL4, and I couldn't get it to work. I'm open to using md for raid, or even LVM, if it supports it. I just want to be able to use raid10, as I can't trust raid5 anymore.


IIRC you had two out of four new disks die? So maybe it would be more accurate to say it's your hardware you don't trust. Raid5 is used without problems by ( I assume ) many, many people, myself included. You could have a raid10 and still lose the whole array if two disks that in the same mirror die at once. I guess no software in the world can really overcome bad hardware. That's why we do backups :)

Anyway, perhaps excersizing /stressing the disks for a few days without error would make you feel more confident about the HDs.


Actually, 2 disks did not die. Due to the fact that it was a new raid 5 array (or for whatever reason), it was rebuilding the array. One of the drives had a media error, and this caused the whole array to be lost.
This is exactly what this article warns about:

http://www.miracleas.com/BAARF/RAID5_versus_RAID10.txt


The article doesn't seem to mention the fact that if a disk in a mirror set dies and the remaining disk within the set starts to have data corruption problems, the mirror will be rebuilt from corrupted data.
While this is true, it's far more likely that there will be a media error (i.e. bad sector), and that the system will notice it. With raid 5, it will just kick out the drive, and you can say bye bye to your data. With raid 10, if it happens on one of the disks in the other set, you don't have a problem, and if it happens to the disk in the same set (not very likely), I'm not sure what the outcome will be, but hopefully it can recover? I have just had a windows drive have a whole bunch of bad sectors and I was still able to boot to windows, and copy most of the data off. I can't imagine Linux being any worse.
I don't know what you can do at this point, though. Perhaps make 2 separate mirrors and rsync them? You could keep copies of changes that way.

I know there is a raid10 personality for md. I saw it in the source code. I see people's boot logs all over the web that say this:

md: linear personality registered as nr 1
md: raid0 personality registered as nr 2
md: raid1 personality registered as nr 3
md: raid10 personality registered as nr 9
md: raid5 personality registered as nr 4

Why does CentOS5 not support the raid10 personality? Do i need to custom compile md? Do I need to custom compile the kernel?
Russ





_______________________________________________
CentOS mailing list
CentOS@xxxxxxxxxx
http://lists.centos.org/mailman/listinfo/centos

[Index of Archives]     [CentOS]     [CentOS Announce]     [CentOS Development]     [CentOS ARM Devel]     [CentOS Docs]     [CentOS Virtualization]     [Carrier Grade Linux]     [Linux Media]     [Asterisk]     [DCCP]     [Netdev]     [Xorg]     [Linux USB]
  Powered by Linux