Re: md RAID with enterprise-class SATA or SAS drives

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 05/10/2012 02:42 PM, Daniel Pocock wrote:
> 
> I think you have to look at the average user's perspective: even most IT
> people don't want to know everything about what goes on in their drives.
>  They just expect stuff to work in a manner they consider `sensible'.
> There is an expectation that if you have RAID you have more safety than
> without RAID.  The idea that a whole array can go down because of
> different sectors failing in each drive seems to violate that expectation.

You absolutely do have more safety, you just might not have as much more
safety as you think.  Modern distributions try hard to automate much of
this setup (e.g. Ubuntu tries to set up mdmon for you when you install
mdadm), but it is not 100%.

Expectations have also changed in the past few years, too, in opposing
ways.  One, hard drive capacities have skyrocketed (Yay!), but error
rate specs have not, so typical users are more likely to encounter UREs.

Two, Linux has gained much more acceptance from home users building
media servers and such, with much more exposure to non-enterprise
components.

Not to excuse the situation--just to explain it.  Coding in this
arena is mostly volunteers, too.

>> Coordinating the drive and the controller timeouts is the *only* way
>> to avoid the URE kickout scenario.
> 
> I really think that is something that needs consideration, as a minimum,
> should md log a warning message if SCTERC is not supported and
> configured in a satisfactory way?

This sounds useful.

>> Changing TLER/ERC when an array becomes degraded for a real hardware
>> failure is a useful idea. I think I'll look at scripting that.
> 
> Ok, so I bought an enterprise grade drive, the WD RE4 (2TB) and I'm
> about to add it in place of the drive that failed.
> 
> I did a quick check with smartctl:
> 
> # smartctl -a /dev/sdb -l scterc
> ....
> SCT Error Recovery Control:
>            Read:     70 (7.0 seconds)
>           Write:     70 (7.0 seconds)
> 
> so the TLER feature appears to be there.  I haven't tried changing it.
> 
> For my old Barracuda 7200.12 that is still working, I see this:
> 
> SCT Error Recovery Control:
>            Read: Disabled
>           Write: Disabled

You should try changing it.  Drives that don't support it won't even
show you that.

You can then put "smartctl -l scterc,70,70 /dev/sdX" in /etc/rc.local or
your distribution's equivalent.

> and a diff between the full output for both drives reveals the following:
> 
> -SCT capabilities:             (0x103f) SCT Status supported.
> +SCT capabilities:             (0x303f) SCT Status supported.
>                                         SCT Error Recovery Control
> supported.
>                                         SCT Feature Control supported.
>                                         SCT Data Table supported.
> 
> 
> 
> 
>>> Here are a few odd things to consider, if you're worried about this topic:
>>>
>>> * Using smartctl to increase the ERC timeout on enterprise SATA
>>> drives, say to 25 seconds, for use with md. I have no idea if this
>>> will cause the drive to actually try different methods of recovery,
>>> but it could be a good middle ground.
>>
> 
> What are the consequences if I don't do that?  I currently have 7
> seconds on my new drive.  If md can't read a sector from the drive, will
> it fail the whole drive?  Will it automatically read the sector from the
> other drive so the application won't know something bad happened?  Will
> it automatically try to re-write the sector on the drive that couldn't
> read it?

MD fails drives on *write* errors.  It reconstructs from mirrors or
parity on read errors and writes the result back to the origin drive.

> Would you know how btrfs behaves in that same scenario - does it try to
> write out the sector to the drive that failed the read?  Does it also
> try to write out the sector when a read came in with a bad checksum and
> it got a good copy from the other drive?

I haven't experimented with btrfs yet.  It is still marked experimental.

Phil
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux