Re: suse_enterprise_storage3_rbd_LIO_vmware_performance_bad

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

my experience:

ceph + iscsi ( multipath ) + vmware == worst

Better you search for another solution.

vmware + nfs + vmware might have a much better performance.

--------

If you are able to get vmware run with iscsi and ceph, i would be
>>very<< intrested in what/how you did that.

-- 
Mit freundlichen Gruessen / Best regards

Oliver Dzombic
IP-Interactive

mailto:info@xxxxxxxxxxxxxxxxx

Anschrift:

IP Interactive UG ( haftungsbeschraenkt )
Zum Sonnenberg 1-3
63571 Gelnhausen

HRB 93402 beim Amtsgericht Hanau
Geschäftsführung: Oliver Dzombic

Steuer Nr.: 35 236 3622 1
UST ID: DE274086107


Am 01.07.2016 um 07:04 schrieb mq:
> Hi list
> I have tested suse enterprise storage3 using 2 iscsi  gateway attached
> to  vmware. The performance is bad.  I have turn off  VAAI following the
> (https://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1033665)
> <https://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1033665%29>.  
> My cluster
> 3 ceph nodes :2*E5-2620 64G , mem 2*1Gbps
> (3*10K SAS, 1*480G  SSD) per node, SSD as journal
> 1 vmware node  2*E5-2620 64G , mem 2*1Gbps
>  
> # ceph -s
>     cluster 0199f68d-a745-4da3-9670-15f2981e7a15
>      health HEALTH_OK
>      monmap e1: 3 mons at
> {node1=192.168.50.91:6789/0,node2=192.168.50.92:6789/0,node3=192.168.50.93:6789/0}
>             election epoch 22, quorum 0,1,2 node1,node2,node3
>      osdmap e200: 9 osds: 9 up, 9 in
>             flags sortbitwise
>       pgmap v1162: 448 pgs, 1 pools, 14337 MB data, 4935 objects
>             18339 MB used, 5005 GB / 5023 GB avail
>                  448 active+clean
>   client io 87438 kB/s wr, 0 op/s rd, 213 op/s wr
>  
> sudo ceph osd tree
> ID WEIGHT  TYPE NAME      UP/DOWN REWEIGHT PRIMARY-AFFINITY
> -1 4.90581 root default
> -2 1.63527     host node1
> 0 0.54509         osd.0       up  1.00000          1.00000
> 1 0.54509         osd.1       up  1.00000          1.00000
> 2 0.54509         osd.2       up  1.00000          1.00000
> -3 1.63527     host node2
> 3 0.54509         osd.3       up  1.00000          1.00000
> 4 0.54509         osd.4       up  1.00000          1.00000
> 5 0.54509         osd.5       up  1.00000          1.00000
> -4 1.63527     host node3
> 6 0.54509         osd.6       up  1.00000          1.00000
> 7 0.54509         osd.7       up  1.00000          1.00000
> 8 0.54509         osd.8       up  1.00000          1.00000
>  
>  
>  
> An linux vm in vmmare, running fio.  4k randwrite result just 64 IOPS
> lantency is high,dd test just 11MB/s.
>  
> fio -ioengine=libaio -bs=4k -direct=1 -thread -rw=randwrite -size=100G
> -filename=/dev/sdb  -name="EBS 4KB randwrite test" -iodepth=32 -runtime=60
> EBS 4KB randwrite test: (g=0): rw=randwrite, bs=4K-4K/4K-4K/4K-4K,
> ioengine=libaio, iodepth=32
> fio-2.0.13
> Starting 1 thread
> Jobs: 1 (f=1): [w] [100.0% done] [0K/131K/0K /s] [0 /32 /0  iops] [eta
> 00m:00s]
> EBS 4KB randwrite test: (groupid=0, jobs=1): err= 0: pid=6766: Wed Jun
> 29 21:28:06 2016
>   write: io=15696KB, bw=264627 B/s, iops=64 , runt= 60737msec
>     slat (usec): min=10 , max=213 , avg=35.54, stdev=16.41
>     clat (msec): min=1 , max=31368 , avg=495.01, stdev=1862.52
>      lat (msec): min=2 , max=31368 , avg=495.04, stdev=1862.52
>     clat percentiles (msec):
>      |  1.00th=[    7],  5.00th=[    8], 10.00th=[    8], 20.00th=[    9],
>      | 30.00th=[    9], 40.00th=[   10], 50.00th=[  198], 60.00th=[  204],
>      | 70.00th=[  208], 80.00th=[  217], 90.00th=[  799], 95.00th=[ 1795],
>      | 99.00th=[ 7177], 99.50th=[12649], 99.90th=[16712], 99.95th=[16712],
>      | 99.99th=[16712]
>     bw (KB/s)  : min=   36, max=11960, per=100.00%, avg=264.77,
> stdev=1110.81
>     lat (msec) : 2=0.03%, 4=0.23%, 10=40.93%, 20=0.48%, 50=0.03%
>     lat (msec) : 100=0.08%, 250=39.55%, 500=5.63%, 750=2.91%, 1000=1.35%
>     lat (msec) : 2000=4.03%, >=2000=4.77%
>   cpu          : usr=0.02%, sys=0.22%, ctx=2973, majf=0,
> minf=18446744073709538907
>   IO depths    : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.2%, 16=0.4%, 32=99.2%,
>>=64=0.0%
>      submit    : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%,
>>=64=0.0%
>      complete  : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.1%, 64=0.0%,
>>=64=0.0%
>      issued    : total=r=0/w=3924/d=0, short=r=0/w=0/d=0
>  
> Run status group 0 (all jobs):
>   WRITE: io=15696KB, aggrb=258KB/s, minb=258KB/s, maxb=258KB/s,
> mint=60737msec, maxt=60737msec
>  
> Disk stats (read/write):
>   sdb: ios=83/3921, merge=0/0, ticks=60/1903085, in_queue=1931694,
> util=100.00%
> 
> anyone can give me some suggestion to improve the performance ?
> 
> Regards
> 
> MQ
> 
> 
> 
> 
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux