Re: raid 1 question

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]



On Thu, Mar 21, 2013 at 10:04 PM, Dave Johansen <davejohansen@xxxxxxxxx>wrote:

> On Tue, Mar 12, 2013 at 5:04 PM, SilverTip257 <silvertip257@xxxxxxxxx>
> wrote:
> >
> > On Tue, Mar 12, 2013 at 1:28 PM, Dave Johansen
> > <davejohansen@xxxxxxxxx>wrote:
> >
> > > On Fri, Mar 8, 2013 at 6:14 AM, SilverTip257 <silvertip257@xxxxxxxxx>
> > > wrote:
> > > >
> > > > On Thu, Mar 7, 2013 at 6:54 PM, Gerry Reno <greno@xxxxxxxxxxx>
> wrote:
> > > >
> > > > > On 03/07/2013 06:52 PM, Les Mikesell wrote:
> > > > > > On Thu, Mar 7, 2013 at 5:40 PM, John R Pierce
> > > > > > <pierce@xxxxxxxxxxxx>
> > > > > wrote:
> > > > > >> On 3/7/2013 3:35 PM, Gerry Reno wrote:
> > > > > >>> Dave,  I've been using software raid with every type of RedHat
> > > distro
> > > > > RH/CentOS/Fedora for over 10 years without any
> > > > > >>> serious difficulties.  I don't quite understand the logic in
> all
> > > these
> > > > > negative statements about software raid on that
> > > > > >>> wiki page.  The worst I get into is I have to boot from a
> > > > > >>> bootdisk
> > > if
> > > > > the MBR gets corrupted for any reason.  No big
> > > > > >>> deal.  Just rerun grub.
> > > > > >> have you been putting /boot on a mdraid?  that's what the
> article
> > > > > >> is
> > > > > >> recommending against.
> > > > > > I've put /boot on md  raid1 on a lot of machines (always drives
> > > > > > small
> > > > > > enough to be MBR based) and never had any problem with the
> > > > > > partition
> > > > > > looking enough like a native one for grub to boot it.  The worst
> > > thing
> > > > >
> > > >
> > > > No problems here either - I have had /boot on software raid1 on quite
> > > > a
> > > few
> > > > systems past and present.
> > > >
> > > >
> > > > > > I've seen about it is that some machines change their idea of
> bios
> > > > >
> > > >
> > > > If I do have a drive fail, I can frequently hot-remove them and
> > > > hot-add
> > > the
> > > > replacement drive to get it resyncing without powering off.
> > > >
> > > >
> > > > > > disk 0 and 1 when the first one fails, so your grub setup might
> be
> > > > > > wrong even after you do it on the 2nd disk - and that would be
> the
> > > > > > same with/without raid.   As long as you are prepared to boot
> from
> > > > > > a
> > > > > > rescue disk you can fix it easily anyway.
> > > > > >
> > > > > Good point, Les.   Rescue disks and bootdisks are key and critical
> > > > > if
> > > > > you're going to use software raid.
> > > > >
> > > > >
> > > > I think we could argue that rescue disks are a necessity regardless
> > > > one
> > > is
> > > > using software raid or not.  :)
> > >
> > > Thanks for all of the helpful info, and now I have a follow on
> > > question. I have a Dell m6500 that I've been running as a RAID 1 using
> > > the BIOS RAID on RHEL 5. The issue is that when you switch one of the
> > > drives, the BIOS renames the RAID and then RHEL 5 doesn't recognize it
> > > anymore. So here are my questions:
> > >
> > > 1) Has this issue of handling the renaming been resolved in RHEL 6?
> > > (my guess is no)
> > >
> >
> > I've not seen any weird bios naming issues.
>
> It's an Intel Software RAID and whenever either of the drives are
> switched, the system won't boot at all because it says it can't find
> the device. I end up having to boot to a rescue disk and then manually
> edit the init image to get it to boot.
>
> >
> >
> > > 2) Would a software RAID be a better choice than using the BIOS RAID?
> > >
> >
> > It depends!  In another thread on this list someone said they prefer the
> > reliable Linux toolset over the manufacturer tools/RAID controllers.
> >
> > In a way it comes down to what you can afford and what you are
> comfortable
> > with.
> > And then there are chips on the hardware RAID controllers on which the
> > RAID
> > operations are offloaded.
> >
> >
> > > 3) If a software RAID is the better choice, are there going to be an
> > > impact on performance/stability/etc?
> > >
> >
> > Supposedly hardware RAID performs better.  I've not done any tests to
> > quantify this statement I've heard others make.
> >
> > Since it's software RAID, the OS will be using a few CPU cycles to handle
> > the softraid.  But I'll doubt you'll miss those CPU cycles ... I haven't
> > and I have a mix of hardware raid and software raid systems.  In the end
> > that it will come down to drive performance in terms of how many IOPS you
> > can squeeze out of your drives.
> >
> > Make certain to set up checks for your array health and disk health
> > (smartd).  If you use hardware raid many controllers don't allow directly
> > accessing the drives with smartctl ... you have to use a vendor
> > binary/utility (or open source one if available).
> >
> > And then there's aligning your hardware RAID boundaries... [0] [1]  :)
> >
> > [0]
> >
> >
> http://www.mysqlperformanceblog.com/2011/06/09/aligning-io-on-a-hard-disk-raid-the-theory/
> > [1]
> >
> >
> http://www.mysqlperformanceblog.com/2011/12/16/setting-up-xfs-the-simple-edition/
>
> Thanks to everyone for all the info and help. It sounds like the best
> solution is a software RAID with /boot being a non-RAID partition and
> then being rsynced on to the second drive, so I'll give that a whirl.
>

Just a bit of emphasis here:

I've had success with /boot being part of a RAID1 software array and
installing GRUB on the MBR of both disks.  That way if the main/first disk
fails the other disk still has GRUB installed and the system will boot if
rebooted.

RAID1 is viewed as single disks by GRUB.  That is it boots off the _first
drive_ -- it is not until the Linux kernel has loaded and mdadm has
assembled the drives that you actually have a RAID1.  Since GRUB only boots
off one drive it is prudent to install GRUB on both disks when set up this
way.

Putting /boot on a RAID1 software array will save you having to rsync /boot
to the other drive.  And booting to a rescue CD to install GRUB on the new
drive after the primary drive died.  The above configuration is some work
up front, but less hassle in the wake of a drive failure.

Try it on a test system if you don't trust the configuration ... it will
boot.  :)
I have this configuration on fifteen or more systems (rough estimate) and
some have been in service for years.


>
> Thanks again,
> Dave
> _______________________________________________
> CentOS mailing list
> CentOS@xxxxxxxxxx
> http://lists.centos.org/mailman/listinfo/centos
>


Best Regards,
-- 
---~~.~~---
Mike
//  SilverTip257  //
_______________________________________________
CentOS mailing list
CentOS@xxxxxxxxxx
http://lists.centos.org/mailman/listinfo/centos


[Index of Archives]     [CentOS]     [CentOS Announce]     [CentOS Development]     [CentOS ARM Devel]     [CentOS Docs]     [CentOS Virtualization]     [Carrier Grade Linux]     [Linux Media]     [Asterisk]     [DCCP]     [Netdev]     [Xorg]     [Linux USB]
  Powered by Linux