Re: Re: Anaconda doesn't support raid10

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]



Chris Croome wrote:
Hi

On Fri 11-May-2007 at 09:38:22AM -0400, Ross S. W. Walker wrote:
The entire use all four disks for /boot makes no sense if two disks
belonging to the same mirror for the lvm go down. Please stop this
nonsense about surviving everything to no benefit. You can have
three disks fail and still have a working /boot. For what?
I think the idea of the 4 partition raid1 was more of, what else is he
going to do with the 200MB at the beginning of each disk which he has
because of partition symmetry across drives?

Makes sense to just dup the partition setup from one to the other and
now with grub and a working /boot on each disk the order of the drives
is no longer important, he can take all 4 out, play 4 disk monty, slap
them back in and the system should come up without a problem.

FWIW this is what I did with the last server I built which had 4x500gb
drives -- a RAID 1 /boot on 4 drives. The trick for this is to edit your
grub.conf so that you can boot off any drive and run grub-install on all
4 drives, also you have to remember to manually edit your grub.conf
after each kernel upgrade to add the 3 extra disks:

 title CentOS (2.6.18-8.1.1.el5xen) Disk 0
         root (hd0,0)
         kernel /xen.gz-2.6.18-8.1.1.el5
         module /vmlinuz-2.6.18-8.1.1.el5xen ro root=/dev/VolGroup00/Root
         module /initrd-2.6.18-8.1.1.el5xen.img
 title CentOS (2.6.18-8.1.1.el5xen) Disk 1
         root (hd1,0)
         kernel /xen.gz-2.6.18-8.1.1.el5
         module /vmlinuz-2.6.18-8.1.1.el5xen ro root=/dev/VolGroup00/Root
         module /initrd-2.6.18-8.1.1.el5xen.img
 title CentOS (2.6.18-8.1.1.el5xen) Disk 2
         root (hd2,0)
         kernel /xen.gz-2.6.18-8.1.1.el5
         module /vmlinuz-2.6.18-8.1.1.el5xen ro root=/dev/VolGroup00/Root
         module /initrd-2.6.18-8.1.1.el5xen.img
 title CentOS (2.6.18-8.1.1.el5xen) Disk 3
         root (hd3,0)
         kernel /xen.gz-2.6.18-8.1.1.el5
         module /vmlinuz-2.6.18-8.1.1.el5xen ro root=/dev/VolGroup00/Root
         module /initrd-2.6.18-8.1.1.el5xen.img

If I had read this thread before I set up this machine I'd have used
RAID 6 from the rest of the space, but I used RAID 5 with a hot spare,
with LVM on top of that.

Before the machine was moved to the colo I tried pulling disks out while
it was running and this worked without a problem, which was nice :-)

Chris



Chris,

I didn't have to install grub on any of the other volumes, and the server seemed to do well after drive failure (I pulled out drives 1 and 3). In my opinion, neither raid5 or raid6 makes sense with 4 drives, as you will get the same amount of space with raid10, but much better performance and availability (although raid6 is supposed to withstand 2 drive failures, in my tests it has not done well, and neither has raid5). As you might've read from the thread, if you want to put the / volume on the raid10, it will not be possible during the install, but you can set up 2 raid1 volumes, and do an LVM stripe across them which should yield comparable performance.
Russ


_______________________________________________
CentOS mailing list
CentOS@xxxxxxxxxx
http://lists.centos.org/mailman/listinfo/centos

[Index of Archives]     [CentOS]     [CentOS Announce]     [CentOS Development]     [CentOS ARM Devel]     [CentOS Docs]     [CentOS Virtualization]     [Carrier Grade Linux]     [Linux Media]     [Asterisk]     [DCCP]     [Netdev]     [Xorg]     [Linux USB]
  Powered by Linux