Re: mdadm raid1 read performance

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu, May 5, 2011 at 3:38 PM, David Brown <david@xxxxxxxxxxxxxxx> wrote:
> On 05/05/2011 12:41, Keld Jørn Simonsen wrote:
>>
>> On Thu, May 05, 2011 at 09:26:45AM +0200, David Brown wrote:
>>>
>>> On 05/05/2011 02:40, Liam Kurmos wrote:
>>>>
>>>> Cheers Roberto,
>>>>
>>>> I've got the gist of the far layout from looking at wikipedia. There
>>>> is some clever stuff going on that i had never considered.
>>>> i'm going for f2 for my system drive.
>>>>
>>>> Liam
>>>>
>>>
>>> For general use, raid10,f2 is often the best choice.  The only
>>> disadvantage is if you have applications that make a lot of synchronised
>>> writes, as writes take longer (everything must be written twice, and
>>> because the data is spread out there is more head movement).  For most
>>> writes this doesn't matter - the OS caches the writes, and the app
>>> continues on its way, so the writes are done when the disks are not
>>> otherwise used.  But if you have synchronous writes, so that the app
>>> will wait for the write to complete, it will be slower (compared to
>>> raid10,n2 or raid10,o2).
>>
>> Yes syncroneous writes would be significantly slower.
>> I have not seen benchmarks on it, tho.
>> Which applications typically use syncroneous IO?
>> Maybe not that many.
>> Do databases do that, eg postgresql and mysql?
>>
>
> Database servers do use synchronous writes (or fsync() calls), but I suspect
> that they won't suffer much if these are slow unless you have a great deal
> of writes - they typically write to the transaction log, fsync(), write to
> the database files, fsync(), then write to the log again and fsync().  But
> they will buffer up their writes as needed in a separate thread or process -
> it should not hinder their read processes.
>
> Lots of other applications also use fsync() whenever they want to be sure
> that data is written to the disk.  A prime example is sqlite, which is used
> by many other programs.  If you have your disk systems and file systems set
> up as a typical home user, there is little problem - the disk write caches
> and file system caches will ensure that the app thinks the write is complete
> long before it hits the disk surfaces anyway (thus negating the whole point
> of using fsync() in the first place...).  But if you have a more paranoid
> setup, so that your databases or other files will not get corrupted by power
> fails or OS crashes, then you have write barriers enabled on the filesystems
> and write caches disabled on the disks.
I guess you mess things a bit - one should disable write cache or
enable barriers at one time, not both. Here goes quote from XFS faq:
"Write barrier support is enabled by default in XFS since kernel
version 2.6.17. It is disabled by mounting the filesystem with
"nobarrier". Barrier support will flush the write back cache at the
appropriate times (such as on XFS log writes). "
http://xfs.org/index.php/XFS_FAQ#Write_barrier_support.

> fsync() will then take time - and it will slow down programs that wait for fsync().
>
> I've not done (or seen) any benchmarks on this, and I don't think it will be
> noticeable to most users.  But it's a typical tradeoff - if you are looking
> for high reliability even with power failures or OS crashes, then you pay
> for it in some kinds of performance.
>
>
>>> The other problem with raid10 layout is booting - bootloaders don't much
>>> like it.  The very latest version of grub, IIRC, can boot from raid10 -
>>> but it can be awkward.  There are lots of how-tos around the web for
>>> booting when you have raid, but by far the easiest is to divide your
>>> disks into partitions:
>>>
>>> sdX1 = 1GB
>>> sdX2 = xGB
>>> sdX3 = yGB
>>>
>>> Put all your sdX1 partitions together as raid1 with metadata layout
>>> 0.90, format as ext3 and use it as /boot.  Any bootloader will work fine
>>> with that (don't forget to install grub on each disk's MBR).
>>>
>>> Put your sdX2 partitions together as raid10,f2 for swap.
>>>
>>> Put the sdX3 partitions together as raid10,f2 for everything else.  The
>>> most flexible choice is to use LVM here and make logical partitions for
>>> /, /home, /usr, etc.  But you can also partition up the md device in
>>> distinct fixed partitions for /, /home, etc. if you want.
>>
>> there is a similar layout of your disks described in
>>
>> https://raid.wiki.kernel.org/index.php/Preventing_against_a_failing_disk
>>
>
> They've stolen my ideas!  Actually, I think this setup is fairly obvious
> when you think through the workings of raid and grub, and it's not
> surprising that more than one person has independently picked the same
> arrangement.
>
>>> Don't try and make sdX3 and sdX4 groups and raids for separate / and
>>> /home (unless you want to use different raid levels for these two
>>> groups).  Your disks are faster near the start (at the outer edge of the
>>> disk), so you get the best speed by making the raid10,f2 from almost the
>>> whole disk.
>>
>> Hmm, I think the root partition actually would have more accesses than
>> /home and other partitions, so it may be beneficial to give the fastest
>> disk sectors to a separate root partition. Comments?
>>
>
> If you make the root logical volume first, then the home logical volume (or
> fixed partitions within the raid), then you will automatically get faster
> access for it.  The arrangement on the disk (for a two disk raid10,far) will
> then be:
>
> Boot1 SwapA1 SwapB2 RootA1 HomeA1 <spareA1> RootB2 HomeB2 <spareB2>
> Boot2 SwapB1 SwapA2 RootB1 HomeB1 <spareB1> RootA2 HomeA2 <spareA2>
>
> Here "A" and "B" are stripes, while "1" and "2" are copies.
>
> <spare> is unallocated LVM space.
>
> Since Boot is very small, it negligible for performance - it doesn't matter
> that it takes the fastest few tracks.  Swap gets as high speed as the disk
> can support.  Then root will be faster than home, but both will still be
> better than the disk's average speed since one copy of the data is within
> the outer half of the disk.
>
>
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>



-- 
Best regards,
[COOLCOLD-RIPN]
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux