Re: suse_enterprise_storage3_rbd_LIO_vmware_performance_bad

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello,

On Fri, 1 Jul 2016 13:04:45 +0800 mq wrote:

> Hi list
> I have tested suse enterprise storage3 using 2 iscsi  gateway attached
> to  vmware. The performance is bad.  

First off, it's somewhat funny that you're testing the repackaged SUSE
Ceph, but asking for help here (with Ceph being owned by Red Hat).

Aside from that, you're not telling us what these 2 iSCSI gateways are
(SW, HW specs/configuration).

Having iSCSI on top of Ceph is by the very nature of things going to be
slower than native Ceph.

Use "rbd bench" or a VM client with RBD to get a base number of what your
Ceph cluster is capable of, this will help identifying where the slowdown
is.

>I have turn off  VAAI following the
> (https://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1033665)
> <https://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1033665)>.
> My cluster 3 ceph nodes :2*E5-2620 64G , mem 2*1Gbps (3*10K SAS, 1*480G
> SSD) per node, SSD as journal 1 vmware node  2*E5-2620 64G , mem 2*1Gbps 

That's a slow (latency wise) network, but not your problem.
What SSD model? 
A 480GB size suggests a consumer model and that would explain a lot.

Check you storage nodes with atop during the fio runs and see if you can
spot a bottleneck.

Christian

> # ceph -s
>     cluster 0199f68d-a745-4da3-9670-15f2981e7a15
>      health HEALTH_OK
>      monmap e1: 3 mons at
> {node1=192.168.50.91:6789/0,node2=192.168.50.92:6789/0,node3=192.168.50.93:6789/0}
> election epoch 22, quorum 0,1,2 node1,node2,node3 osdmap e200: 9 osds: 9
> up, 9 in flags sortbitwise
>       pgmap v1162: 448 pgs, 1 pools, 14337 MB data, 4935 objects
>             18339 MB used, 5005 GB / 5023 GB avail
>                  448 active+clean
>   client io 87438 kB/s wr, 0 op/s rd, 213 op/s wr
>  
> sudo ceph osd tree
> ID WEIGHT  TYPE NAME      UP/DOWN REWEIGHT PRIMARY-AFFINITY
> -1 4.90581 root default
> -2 1.63527     host node1
> 0 0.54509         osd.0       up  1.00000          1.00000
> 1 0.54509         osd.1       up  1.00000          1.00000
> 2 0.54509         osd.2       up  1.00000          1.00000
> -3 1.63527     host node2
> 3 0.54509         osd.3       up  1.00000          1.00000
> 4 0.54509         osd.4       up  1.00000          1.00000
> 5 0.54509         osd.5       up  1.00000          1.00000
> -4 1.63527     host node3
> 6 0.54509         osd.6       up  1.00000          1.00000
> 7 0.54509         osd.7       up  1.00000          1.00000
> 8 0.54509         osd.8       up  1.00000          1.00000
>  
>  
>  
> An linux vm in vmmare, running fio.  4k randwrite result just 64 IOPS
> lantency is high,dd test just 11MB/s. 
> fio -ioengine=libaio -bs=4k -direct=1 -thread -rw=randwrite -size=100G
> -filename=/dev/sdb  -name="EBS 4KB randwrite test" -iodepth=32
> -runtime=60 EBS 4KB randwrite test: (g=0): rw=randwrite,
> bs=4K-4K/4K-4K/4K-4K, ioengine=libaio, iodepth=32 fio-2.0.13 Starting 1
> thread Jobs: 1 (f=1): [w] [100.0% done] [0K/131K/0K /s] [0 /32 /0  iops]
> [eta 00m:00s] EBS 4KB randwrite test: (groupid=0, jobs=1): err= 0:
> pid=6766: Wed Jun 29 21:28:06 2016 write: io=15696KB, bw=264627 B/s,
> iops=64 , runt= 60737msec slat (usec): min=10 , max=213 , avg=35.54,
> stdev=16.41 clat (msec): min=1 , max=31368 , avg=495.01, stdev=1862.52
>      lat (msec): min=2 , max=31368 , avg=495.04, stdev=1862.52
>     clat percentiles (msec):
>      |  1.00th=[    7],  5.00th=[    8], 10.00th=[    8],
> 20.00th=[    9], | 30.00th=[    9], 40.00th=[   10], 50.00th=[  198],
> 60.00th=[  204], | 70.00th=[  208], 80.00th=[  217], 90.00th=[  799],
> 95.00th=[ 1795], | 99.00th=[ 7177], 99.50th=[12649], 99.90th=[16712],
> 99.95th=[16712], | 99.99th=[16712]
>     bw (KB/s)  : min=   36, max=11960, per=100.00%, avg=264.77,
> stdev=1110.81 lat (msec) : 2=0.03%, 4=0.23%, 10=40.93%, 20=0.48%,
> 50=0.03% lat (msec) : 100=0.08%, 250=39.55%, 500=5.63%, 750=2.91%,
> 1000=1.35% lat (msec) : 2000=4.03%, >=2000=4.77%
>   cpu          : usr=0.02%, sys=0.22%, ctx=2973, majf=0,
> minf=18446744073709538907 IO depths    : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.2%,
> 16=0.4%, 32=99.2%, >=64=0.0% submit    : 0=0.0%, 4=100.0%, 8=0.0%,
> 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% complete  : 0=0.0%, 4=100.0%,
> 8=0.0%, 16=0.0%, 32=0.1%, 64=0.0%, >=64=0.0% issued    :
> total=r=0/w=3924/d=0, short=r=0/w=0/d=0 
> Run status group 0 (all jobs):
>   WRITE: io=15696KB, aggrb=258KB/s, minb=258KB/s, maxb=258KB/s,
> mint=60737msec, maxt=60737msec 
> Disk stats (read/write):
>   sdb: ios=83/3921, merge=0/0, ticks=60/1903085, in_queue=1931694,
> util=100.00%
> 
> anyone can give me some suggestion to improve the performance ?
> 
> Regards
> 
> MQ
> 
> 


-- 
Christian Balzer        Network/Systems Engineer                
chibi@xxxxxxx   	Global OnLine Japan/Rakuten Communications
http://www.gol.com/
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux