No subject

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Latencies:
dfs01: 402 microseconds
dfs02: 322 microseconds
dfs03: 445 microseconds
dfs04: 378 microseconds

Bandwidths:
dfs01: 54 MB/s
dfs02: 62.5 MB/s
dfs03: 64 MB/s
dfs04: 91.5 MB/s

Every server machine has just 1 EBS drive, ext3 filesystem,
2.6.18-xenU-ec2-v1.0 - CFQ IO scheduler.

Any ideas? From the numbers above - does it have any sense to try to
make sw RAID0 with mdadm, or eventually use another filesystem?

Thank you for help.
Regards Karol

On Tue, Mar 22, 2011 at 6:08 PM, Mohit Anchlia <mohitanchlia at gmail.com> wro=
te:
> Can you first run some test with no replica and see what results you
> get? Also, can you look at network latency from client to each of your
> 4 servers and post the results?
>
> On Mon, Mar 21, 2011 at 1:27 AM, karol skocik <karol.skocik at gmail.com> wr=
ote:
>> Hi,
>> =A0I am in the process of evaluation of Gluster for major BI company,
>> but I was surprised by very small write performance on Amazon EBS.
>> Our setup is Gluster 3.1.2, distributed replica 2x2 on 64-bit m1.large
>> instances. Every server node has 1 EBS volume attached to it.
>> The configuration of the distributed replica is a default one, my
>> small attemps to improve performance (io-threads, disabled io-stats
>> and latency-measurement):
>>
>> volume EBSVolume-posix
>> =A0 =A0type storage/posix
>> =A0 =A0option directory /mnt/ebs
>> end-volume
>>
>> volume EBSVolume-access-control
>> =A0 =A0type features/access-control
>> =A0 =A0subvolumes EBSVolume-posix
>> end-volume
>>
>> volume EBSVolume-locks
>> =A0 =A0type features/locks
>> =A0 =A0subvolumes EBSVolume-access-control
>> end-volume
>>
>> volume EBSVolume-io-threads
>> =A0 =A0type performance/io-threads
>> =A0 =A0option thread-count 4
>> =A0 =A0subvolumes EBSVolume-locks
>> end-volume
>>
>> volume /mnt/ebs
>> =A0 =A0type debug/io-stats
>> =A0 =A0option log-level NONE
>> =A0 =A0option latency-measurement off
>> =A0 =A0subvolumes EBSVolume-io-threads
>> end-volume
>>
>> volume EBSVolume-server
>> =A0 =A0type protocol/server
>> =A0 =A0option transport-type tcp
>> =A0 =A0option auth.addr./mnt/ebs.allow *
>> =A0 =A0subvolumes /mnt/ebs
>> end-volume
>>
>> In our test, all clients starts writing to different 1GB file at the sam=
e time.
>> The measured write bandwidth, with 2x2 servers:
>>
>> 1 client: 6.5 MB/s
>> 2 clients: 4.1 MB/s
>> 3 clients: 2.4 MB/s
>> 4 clients: 4.3 MB/s
>>
>> This is not acceptable for our needs. With PVFS2 (I know it's
>> stripping which is very different from replica) we can get up to 35
>> MB/s.
>> 2-3 times slower than that would be understandable. But 5-15 times
>> slower is not, and I would like to know whether there is something we
>> could try out.
>>
>> Could anybody publish their write speeds on similar setup, and tips
>> how to achieve better performance?
>>
>> Thank you,
>> =A0Karol
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>>
>


[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux