On Thu Jan 10, 2013, Chris Murphy wrote: > On Jan 10, 2013, at 3:49 AM, Thomas Fjellstrom <thomas@xxxxxxxxxxxxx> wrote: > > A lot of it will be streaming. Some may end up being random read/writes. > > The test is just to gauge over all performance of the setup. 600MBs read > > is far more than I need, but having writes at 1/3 that seems odd to me. > > Tell us how many disks there are, and what the chunk size is. It could be > too small if you have too few disks which results in a small full stripe > size for a video context. If you're using the default, it could be too big > and you're getting a lot of RWM. Stan, and others, can better answer this. As stated earlier, its a 7x2TB array. > You said these are unpartitioned disks, I think. In which case alignment of > 4096 byte sectors isn't a factor if these are AF disks. They are AF disks. > Unlikely to make up the difference is the scheduler. Parallel fs's like XFS > don't perform nearly as well with CFQ, so you should have a kernel > parameter elevator=noop. > > Another thing to look at is md/stripe_cache_size which probably needs to be > higher for your application. I'll look into it. > Another thing to look at is if you're using XFS, what your mount options > are. Invariably with an array of this size you need to be mounting with > the inode64 option. I'm not sure, but I think that's the default. > > The reason I've selected RAID6 to begin with is I've read (on this > > mailing list, and on some hardware tech sites) that even with SAS > > drives, the rebuild/resync time on a large array using large disks > > (2TB+) is long enough that it gives more than enough time for another > > disk to hit a random read error, > > This is true for high density consumer SATA drives. It's not nearly as > applicable for low to moderate density nearline SATA which has an order of > magnitude lower UER, or for enterprise SAS (and some enterprise SATA) > which has yet another order of magnitude lower UER. So it depends on the > disks, and the RAID size, and the backup/restore strategy. Plain old seagate baracudas, so not the best but at least they aren't greens. > Another way people get into trouble with the event you're talking about, is > they don't do regular scrubs or poll drive SMART data. I have no empirical > data, but I'd expect much better than order of magnitude lower array loss > during a rebuild when the array is being properly maintained, rather than > considering it a push button "it's magic" appliance to be forgotten about. Debian seems to set up a weekly or monthly scrub, which I leave on due to reading that same fact. > > Chris Murphy-- > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html -- Thomas Fjellstrom thomas@xxxxxxxxxxxxx -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html