Re: Incredibly poor performance of mdraid-1 with 2 SSD Samsung 840 PRO

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Ben = Andrei, sorry for the typo.

2013/4/22 Tommy Apel <tommyapeldk@xxxxxxxxx>:
> Stan>
> That was exactly what I was trying to show, that you result may vary
> depending on data and backing device, as far as the raid1 goes it
> doesn't care much for the data beeing passed through it.
>
> Ben>
> could you try to run iostat -x 2 for a few minuts just to make sure
> there is no other I/O going on the device before running your tests,
> and then run the tests with fio instead of dd ?
>
> fio write test > fio --rw=write --filename=testfile --bs=1048576
> --size=4294967296 --ioengine=psync --end_fsync=1 --invalidate=1
> --direct=1 --name=writeperftest
>
> /Tommy
>
> 2013/4/22 Stan Hoeppner <stan@xxxxxxxxxxxxxxxxx>:
>> On 4/21/2013 2:56 PM, Tommy Apel wrote:
>>> Calm the f. down, I was just handing over some information, sorry your
>>> day was ruined mr. high and mighty, use the info for whatever you want
>>> to but flaming me is't going to help anyone.
>>
>> Your tantrum aside, the Intel 330, as well as all current Intel SSDs,
>> uses the SandForce 2281 controller.  The SF2xxx series' write
>> performance is limited by the compressibility of the data.  What you're
>> doing below is simply showcasing the write bandwidth limitation of the
>> SF2xxx controllers with incompressible data.
>>
>> This is not relevant to md.  And it's not relevant to Andrei.  It turns
>> out that the Samsung 840 SSDs have consistent throughput because they
>> don't rely on compression.
>>
>> --
>> Stan
>>
>>
>>> 2013/4/21 Stan Hoeppner <stan@xxxxxxxxxxxxxxxxx>:
>>>> On 4/21/2013 7:23 AM, Tommy Apel wrote:
>>>>> Hello, FYI I'm getting ~68MB/s on two intel330 in RAID1 aswell on
>>>>> vanilla 3.8.8 and 3.9.0-rc3 when writing random data and ~236MB/s
>>>>> writing from /dev/zero
>>>>>
>>>>> mdadm -C /dev/md0 -l 1 -n 2 --assume-clean --force --run /dev/sdb /dev/sdc
>>>>
>>>>
>>>>> openssl enc -aes-128-ctr -pass pass:"$(dd if=/dev/urandom bs=128
>>>>> count=1 2>/dev/null | base64)" -nosalt < /dev/zero | pv -pterb >
>>>>> /run/fill ~1.06GB/s
>>>>
>>>> What's the purpose of all of this?  Surely not simply to create random
>>>> data, which is accomplished much more easily.  Are you sand bagging us
>>>> here with a known bug, or simply trying to show off your mad skillz?
>>>> Either way this is entirely unnecessary for troubleshooting an IO
>>>> performance issue.  dd doesn't (shouldn't) care if the bits are random
>>>> or not, though the Intel SSD controller might, as well as other layers
>>>> you may have in your IO stack.  Keep it simple so we can isolate one
>>>> layer at a time.
>>>>
>>>>> dd if=/run/fill of=/dev/null bs=1M count=1024 iflag=fullblock ~5.7GB/s
>>>>> dd if=/run/fill of=/dev/md0 bs=1M count=1024 oflag=direct ~68MB/s
>>>>> dd if=/dev/zero of=/dev/md0 bs=1M count=1024 oflag=direct ~236MB/s
>>>>
>>>> Noting the above, it's interesting that you omitted this test
>>>>
>>>>   dd if=/run/fill of=/dev/sdb bs=1M count=1024 oflag=direct
>>>>
>>>> preventing an apples to apples comparison between raw SSD device and
>>>> md/RAID1 performance with your uber random file as input.
>>>>
>>>> --
>>>> Stan
>>>>
>>> --
>>> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
>>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>
>>
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux