Re: Re: Raid5 issues
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Toby Bluhm wrote:
Ruslan Sivak wrote:
Toby Bluhm wrote:
Ruslan Sivak wrote:
Feizhou wrote:
Ruslan Sivak wrote:
Feizhou wrote:
I do have a SIL3114 chipset, and I think it's supposed to be
supported by device mapper. When I go to rescue mode, I see it
loading the driver for SIL3112, but nothing appears under
/dev/mapper except control. Are there instructions somewhere
on getting it to use my controller's raid?
Your controller only has a bios chip. It has no raid processing
capability at all.
You need to use mdadm. anaconda should be able to let you create
to mirrors and then create a third array that stripes those md
devices,
_______________________________________________
Anaconda doesn't let me create a stripe raid set on top of a
mirror set. And it doesn't detect it when I do it manually.
Also the bios chip presents additional issues. I believe when I
don't have a raid array set up, it won't boot at all. When I
have it on raid10, I had trouble booting, and when I have it on
concatenation, everything works fine, until a drive is replaced.
At that point, i have to recreate the array, as concatenation is
not a fault tolerant set, and at this point I seem to lose all my
data.
It won't boot at all without a raid array setup? That sounds
really funny.
Actually I'm not 100% sure on this, but I think this is the case.
I believe the first time I set it up as a raid10, assuming that
linux will just ignore it. I installed centos by putting boot on a
raid1, and root on LVM over 2 raid1 sets. I had trouble getting it
to boot.
Is there a way to get it to use the raid that's part of the bios
chip?
Repeat after me. There is no raid that is part of the bios chip.
It is just a simple table.
Yes, I know this is fakeraid, aka softraid, but I was hoping that
using the drivers would make it easier to support raid 10 then with
mdadm, which seems to be impossible to get to work with the
installer. I'm not even sure why the raid10 personality is not
loaded, as it seems to have been part of the mdadm since version 1.7.
Something about device mapper?
You need the fake raid driver dmraid if you are going to set up
stuff in the bios. What version of centos are you trying to
install? libata in Centos 5 should support this without having to
resort to the ide drivers.
_________________________________
I'm trying to install centos 5 - the latest. How would I go about
using dmraid and/or libata? The installer picks up the drives as
individual drives. There is a drive on the silicon image website,
but it's for RHEL4, and I couldn't get it to work. I'm open to
using md for raid, or even LVM, if it supports it. I just want to
be able to use raid10, as I can't trust raid5 anymore.
IIRC you had two out of four new disks die? So maybe it would be
more accurate to say it's your hardware you don't trust. Raid5 is
used without problems by ( I assume ) many, many people, myself
included. You could have a raid10 and still lose the whole array if
two disks that in the same mirror die at once. I guess no software
in the world can really overcome bad hardware. That's why we do
backups :)
Anyway, perhaps excersizing /stressing the disks for a few days
without error would make you feel more confident about the HDs.
Actually, 2 disks did not die. Due to the fact that it was a new
raid 5 array (or for whatever reason), it was rebuilding the array.
One of the drives had a media error, and this caused the whole array
to be lost.
This is exactly what this article warns about:
http://www.miracleas.com/BAARF/RAID5_versus_RAID10.txt
The article doesn't seem to mention the fact that if a disk in a
mirror set dies and the remaining disk within the set starts to have
data corruption problems, the mirror will be rebuilt from corrupted data.
While this is true, it's far more likely that there will be a media
error (i.e. bad sector), and that the system will notice it. With raid
5, it will just kick out the drive, and you can say bye bye to your
data. With raid 10, if it happens on one of the disks in the other set,
you don't have a problem, and if it happens to the disk in the same set
(not very likely), I'm not sure what the outcome will be, but hopefully
it can recover? I have just had a windows drive have a whole bunch of
bad sectors and I was still able to boot to windows, and copy most of
the data off. I can't imagine Linux being any worse.
I don't know what you can do at this point, though. Perhaps make 2
separate mirrors and rsync them? You could keep copies of changes
that way.
I know there is a raid10 personality for md. I saw it in the source
code. I see people's boot logs all over the web that say this:
md: linear personality registered as nr 1
md: raid0 personality registered as nr 2
md: raid1 personality registered as nr 3
md: raid10 personality registered as nr 9
md: raid5 personality registered as nr 4
Why does CentOS5 not support the raid10 personality? Do i need to custom compile md? Do I need to custom compile the kernel?
Russ
_______________________________________________
CentOS mailing list
CentOS@xxxxxxxxxx
http://lists.centos.org/mailman/listinfo/centos
[Index of Archives]
[CentOS]
[CentOS Announce]
[CentOS Development]
[CentOS ARM Devel]
[CentOS Docs]
[CentOS Virtualization]
[Carrier Grade Linux]
[Linux Media]
[Asterisk]
[DCCP]
[Netdev]
[Xorg]
[Linux USB]
|