Re: very strange raid10,f2 performance

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Fri, Apr 30, 2010 at 5:46 PM, Keld Simonsen <keld@xxxxxxxxxx> wrote:
> On Wed, Apr 21, 2010 at 12:02:46PM -0500, Jon Nelson wrote:
>> I was helping somebody else diagnose some issues, and decided to run
>> comparitive tests on my own raid (raid10,f2).
>>
>> The raid10,f2 (md0) is the only physical device backing a volume
>> group, which is then carved into a bunch of (primarily) ext4
>> filesystems.
>> The kernel is 2.6.31.12 (openSUSE) on a Quad Processor AMD Phenom 9150e system.
>> The raid is two Western Digital Caviar Blue drives (WDC WD5000AAKS-00V1A0).
>>
>> The problem: really, really bad I/O performance under certain circumstances.
>>
>> When using an internal bitmap and *synchronous* I/O, applications like
>> dd report 700-800 kB/s.
>> When not using a bitmap at all, and synchronous I/O, dd reports 2.5
>> MB/s (but dstat shows 14MB/s?)
>> Without a bitmap and async I/O (but with fdatasync) I get 65MB/s.
>> *With* a bitmap and using async. I/O (but with fdatasync) I get more
>> like 65MB/s.
>>
>> The system has 3GB of memory and I'm testing with dd if=/dev/zero
>> of=somefile bs=4k count=524288.
>>
>> I'm trying to understand why the synchronous I/O is so bad, but even
>> so I was hoping for more. 65MB/s seems *reasonable* given the
>> raid10,f2 configuration and all of the seeking that such a
>> configuration involves (when writing).
>>
>> The other very strange thing is that the I/O patterns seem very
>> strange. I'll see 14MB/s very consistently as reported by dstat
>> (14MB/s for each sda, sdb, and md0) for 10-15 seconds and then I'll
>> see it drop, sometimes to just 3 or 4 MB/s, for another 10 seconds,
>> and then the pattern repeats.  What's going on here? With absolutely
>> no other load on the system, I would have expected to see something
>> much more consistent.
>
> Hmm, not much response to this.
> The only idea I have for now is misalignment between raid and LVM boundaries.

These aren't 4K disks (as far as I know), so I'm not sure what you
mean by alignment issues.
Using 255 heads, 63 sectors per track:

/dev/sda1 starts on sector 63 and ends on sector 7807589
/dev/sda2 starts on sector 11711385 and ends on sector 482528340

/dev/sdb is partitioned the same
/dev/sda2 and /dev/sdb2 form the raid10,f2.

> Were your dd's done on the raw devices, or via a file system?

Raw (logical) devices carved out of the volume group.

-- 
Jon
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux