RE: RAID 5 - One drive dropped while replacing another

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



> >        That's pretty conservative, yes, for middle of the road
> > availability.  For a system whose necessary availability is not too
> high, it
> > is considerable overkill.  For a system whose availability is critical,
> it's
> > not conservative enough.
> 
> So maybe for my "money-poor environment that could live with a day or
> two of downtime" I'll add a drive or two as my own personal rule of
> thumb. Thanks for the feedback.

	Yeah, as long as you have good backups for all critical data and you
can live with the thought of being at least partially down for an extended
period while the data restores from your backups, then whatever level of
redundancy you can afford should be fine.  If it is critical you be able to
have some portion of the data immediately available at all times, then a
more aggressive (and probably expensive) solution is in order.  You might
also have to give up some performance for greater reliability.

> >        That assumes the RAID1 array elements only have 2 members.  With
> 3
> > members, the reliability goes way up.  Of course, so does the cost.
> 
> Prohibitively so for my situation, at least for the large storage
> volumes.

	I can relate.

> My OS boot partitions are replicated on every drive, so some
> of them have 20+ members, but at 10GB per 2TB, not expensive  8-)

	That really sounds like overkill.  I'm a big fan of keeping my data
storage completely separate from my boot media.  Indeed, the boot drives are
internal to the servers, while the data drives reside in RAID chassis.
Small drives are really cheap (in my case, free), so I simply chunked a
couple of old drives into each server chassis, partitioned them into root,
boot, and swap partitions, and bound them into three RAID1 arrays.  Every
few days, I back up the files on the / and /boot arrays onto the data array
using tar.  Rsync would also be a good candidate.  The space utilization, as
you say, is pretty trivial, as is the drain on the server resources.

> >> depending on luck, whereas RAID6 would allow **any** two (and
> >> RAID6+spare any *three*) drives to fail without my losing data. So I
> >
> >        That's specious.  RAID6 + spare only allows two overlapping
> failures.
> 
> Well yes, but my environment doesn't have pager notification to a
> "hot-spare sysadmin" standing by ready to jump in. In fact the
> replacement hardware would need to be requested/purchase-ordered etc,
> so in that case the spare does make a difference to resilience doesn't
> it.

	It improves availability, not really resilience.  It also of course
impacts reliability.

> If I had the replacement drive handy I'd just make it a hot spare
> rather than keeping it on the shelf anyway.

	Oh, yeah, I'm not disparaging the hot spare.  It's just if two
members suffer overlapping failures, then the array is without any
redundancy until the resync is complete.

> >> On my lower-end systems, a RAID6 over 2TB drives takes about 10-11
> >> hours per failed disk to rebuild, and that's using embedded bitmaps
> >> and with nothing else going on.
> >
> >        I've never had one rebuild from a bare drive that fast.
> 
> This wasn't a bare drive, but a re-add after I'd been doing some grub2
> and maintenance work on another array from SystemRescueCD, not sure
> why the member failed.

	That's a very different matter.  A re-add of a failed member can be
very brief.  When the system has to read every last byte of data from the
live array, calculate parity, and then write the calculated data back to the
blank drive, it can easily take close to or even more than a day per TB,
depending on the system load.

> It's not a particularly fast platform, consumer
> SATA2 Hitachi drives attached to mobo Intel controller, ICH7 I
> believe. cat /proc/mdstat was showing around 60k, while the RAID1s
> rebuild at around double that.
> 
> Would the fact that it was at less than 30% capacity make a difference?

	That, I'm not sure.  I'm not certain of the mechanics of mdadm, and
I have never done a drive replacement on a system that lightly filled.

--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux