Re: max number of devices in raid6 array

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



"Guy Watkins" <linux-raid@xxxxxxxxxxxxxxxx> writes:

> } -----Original Message-----
> } From: linux-raid-owner@xxxxxxxxxxxxxxx [mailto:linux-raid-
> } owner@xxxxxxxxxxxxxxx] On Behalf Of Goswin von Brederlow
> } Sent: Wednesday, August 12, 2009 10:52 PM
> } To: John Robinson
> } Cc: Goswin von Brederlow; David Cure; linux-raid@xxxxxxxxxxxxxxx
> } Subject: Re: max number of devices in raid6 array
> } 
> } "John Robinson" <john.robinson@xxxxxxxxxxxxxxxx> writes:
> } 
> } > On Wed, 12 August, 2009 3:53 pm, Goswin von Brederlow wrote:
> } > [...]
> } >> And compute the overall MTBFS. With how many devices does the MTBFS of
> } a
> } > raid6 drop below that of a single disk?
> } >
> } > First up, we probably want to be talking about Mean Time To Data Loss.
> } > It'll vary enormously depending on how fast you think you can replace
> } dead
> } > drives, which in turn depends on how long a rebuild takes (since a dead
> } > drive doesn't count as having been replaced until the new drive is fully
> } > sync'ed). And building an array that big, it's going to be hard to get
> } > drives all from different batches.
> } >
> } > Anyway, someone asked Google a similar question:
> } > http://answers.google.com/answers/threadview/id/730165.html and the
> } MTTDL
> } > for an 11-disc RAID-5 with 100,000-hour drives and a 24-hour
> } > replacement+rebuild turnaround was 3.8 million hours (433 years), and a
> } > RAID-6 was said to be "hundreds of times" more reliable. The 433 years
> } > figure will be assuming that one drive failure doesn't cause another
> } one,
> } > though, so it's to be taken with a pinch of salt.
> } >
> } > Cheers,
> } >
> } > John.
> } 
> } I would take that with a verry large pinch of salt. From the little
> } experience I have that value doesn't reflects reality.
> } 
> } Unfortunately the MTBFS values for disks vendors give are pretty much
> } totaly dreamed up. So the 100,000-hours for a single drive already has
> } a huge uncertainty. Shouldn't affect the cut of point where the MTBFS
> } of tha raid is less than a single disk though.
> } 
> } Secondly disk failures in a raid are not unrelated. The disk all age
> } and most people don't rotate in new disk regulary. The chance of a
> } disk failure is not uniform over time.
> } 
> } On top of that the stress of rebuilding usualy greatly increases the
> } chances. And with large raids and todays large disks we are talking
> } days to weeks or rebuild time. As you said, the 433 years are assuming
> } that one drive failure doesn't cause another one to fail. In reality
> } that seems to be a real factor though.
> } 
> } 
> } If I understood the math in the URL right then the chance of a disk
> } failing within a week is:
> } 
> } 168/100000 = 0.00168
> } 
> } The chance of 2 disks failing within a week with 25 disks would be:
> } 
> } (1-(1-168/100000)^25)^2 =  ~0.00169448195081717874
> } 
> } The chance of 3 disks failing within a week with 75 disks would be:
> } 
> } (1-(1-168/100000)^75)^3 =  ~0.00166310371815668874
> } 
> } So the cut off values are roughly 25 and 75 disks for raid 5/6. Right?
> } 
> } 
> } Now lets assume, and I'm totally guessing here, the failure is 4 times
> } more likely during a rebuild:
> } 
> } (1-(1-168/100000*4)^7)^2  = ~0.00212541503635
> } (1-(1-168/100000*4)^19)^3 = ~0.00173857193240
> } (1-(1-336/100000*4)^10)^3 = ~0.00202697761277 (two weeks rebuild time)
> } 
> } So cut off is 7 and 19 (10 for 2 week rebuild) disks. Or am I totaly
> } doing the wrong math?
> } 
> } MfG
> }         Goswin
>
> I don't believe a block read error is considered in the MTBF.  A current 2TB
> disk has a "<1 in 10^15" "Non-recoverable read errors per bits read".  That
> is about 1 error per 114 TB read (10^15/8/1024/1024/1024/1024).  So, you
> should get 1 failure per about 114 TB read.  If you had 57 2TB disks + 1
> parity, your chance of a read error should be 1 during a recovery.  If you
> had 29 2TB disks and 1 parity, you should have about 1 failure per 2
> recoveries.  With 6 2TB disks and 1 parity, you should have about 1 failure
> per 10 recoveries. This assumes you had no other disk reads to increase the
> failure rate.
>
> I got the 10^15 from here:
> http://www.wdc.com/en/library/sata/2879-701229.pdf
>
> I hope my math is correct!
>
> Guy

But does that cause data loss? If only one disk has failed on a raid6
a read error is still correctable and rewriting would remap the block
internally in the drive and avoid having to fail the drive. The kernel
does that nowadays or not?

MfG
        Goswin
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux