Re: Optimize RAID0 for max IOPS?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, Jan 26, 2011 at 12:38 PM, Christoph Hellwig <hch@xxxxxxxxxxxxx> wrote:
> On Wed, Jan 26, 2011 at 08:16:16AM +0100, Wolfgang Denk wrote:
>> I will not have a single file system, but several, so I'd probably go
>> with LVM. But - when I then create a LV, eventually smaller than any
>> of the disks, will the data (and thus the traffic) be really distri-
>> buted over all drives, or will I not basicly see the same results as
>> when using a single drive?
>
> Think about it:  if you're doing small IOPs, they usually are smaller
> than the stripe size and you will hit only one disk anyway.  But with
> a raid0 which disk you hit is relatively unpredictable.  With a
> concatentation aligned to the AGs XFS will distribute processes writing
> data to the different AGs and thus the different disks, and you can
> reliably get performance out of them.
>
> If you have multiple filesystems the setup depends a lot on the
> workloads you plan to put on the filesystems.  If all of the filesystems
> on it are busy at the same time just assigning disks to filesystems
> probably gives you the best performace.  If they are busy at different
> times, or some are not busy at all you first want to partition the disk
> into areas for each filesystem and then concatenate them into volumes
> for each filesystem.
>
>
>> [[Note: Block write: drop to 60%, Block read drops to <50%]]
>
> How is the cpu load?  delaylog trades I/O operations for cpu
> utilization.  Together with a raid6, which apparently is the system you
> use here i might overload your system.
>
> And btw, in future please state you have numbers for a totally different
> setup then the one you're asking questions for.  Comparing a raid6 setup
> to striping/concatenation is completely irrelevant.
>
>>
>> [[Add nobarriers]]
>>
>> # mount -o remount,nobarriers /mnt/tmp
>> # mount | grep /mnt/tmp
>> /dev/mapper/castor0-test on /mnt/tmp type xfs (rw,noatime,delaylog,logbsize=262144,nobarriers)
>
>  a) the option is called nobarrier
>  b) it looks like your mount implementation is really buggy as it shows
>    random options that weren't actually parsed and accepted by the
>    filesystem.
cat /proc/mounts may help i guess

>
>> [[Again, degradation of about 10% for block read; with only minod
>> advantages for seq. delete and random create]]
>
> I really don't trust the numbers.  nobarrier sends down less I/O
> requests, and avoids all kinds of queue stalls.  How repetable are these
> benchmarks?  Do you also see it using a less hacky benchmark than
> bonnier++?
>
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>



-- 
Best regards,
[COOLCOLD-RIPN]
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux