Re: full ssd setup preliminary hammer bench

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



>>Any quick write performance data? 

4k randwrite
iops : 12K
hostcpu: 85.5 idle
client cpu : 98,5 id idle
disk util : 100%  (this is the bottleneck).

This s3500 drives can do around 25K rand 4K write with O_DSYNC 

So, with ceph double write (journal + datas), that's explain the 12K





----- Mail original -----
De: "Michal Kozanecki" <mkozanecki@xxxxxxxxxx>
À: "aderumier" <aderumier@xxxxxxxxx>, "Mark Nelson" <mnelson@xxxxxxxxxx>, "ceph-users" <ceph-users@xxxxxxxxxxxxxx>
Envoyé: Vendredi 17 Avril 2015 18:13:59
Objet: RE: full ssd setup preliminary hammer bench

Any quick write performance data? 

Michal Kozanecki | Linux Administrator | E: mkozanecki@xxxxxxxxxx 


-----Original Message----- 
From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf Of Alexandre DERUMIER 
Sent: April-17-15 11:38 AM 
To: Mark Nelson; ceph-users 
Subject:  full ssd setup preliminary hammer bench 

Hi Mark, 

I finally got my hardware for my production full ssd cluster. 

Here a first preliminary bench. (1osd). 

I got around 45K iops with randread 4K with a small 10GB rbd volume 


I'm pretty happy because I don't see anymore huge cpu difference between krbd && lirbd. 
In my previous bench I was using debian wheezy as client, now it's a centos 7.1, so maybe something is different (glibc,...). 

I'm planning to do big benchmark centos vs ubuntu vs debian, client && server, to compare. 
I have 18 osd ssd for the benchmarks. 







results : rand 4K : 1 osd 
------------------------- 

fio + librbd: 
------------ 
iops: 45.1K 

clat percentiles (usec): 
| 1.00th=[ 358], 5.00th=[ 406], 10.00th=[ 446], 20.00th=[ 556], 
| 30.00th=[ 676], 40.00th=[ 1048], 50.00th=[ 1192], 60.00th=[ 1304], 
| 70.00th=[ 1400], 80.00th=[ 1496], 90.00th=[ 1624], 95.00th=[ 1720], 
| 99.00th=[ 1880], 99.50th=[ 1928], 99.90th=[ 2064], 99.95th=[ 2128], 
| 99.99th=[ 2512] 

cpu server : 89.1 iddle 
cpu client : 92,5 idle 

fio + krbd 
---------- 
iops:47.5K 

clat percentiles (usec): 
| 1.00th=[ 620], 5.00th=[ 636], 10.00th=[ 644], 20.00th=[ 652], 
| 30.00th=[ 668], 40.00th=[ 676], 50.00th=[ 684], 60.00th=[ 692], 
| 70.00th=[ 708], 80.00th=[ 724], 90.00th=[ 756], 95.00th=[ 820], 
| 99.00th=[ 1004], 99.50th=[ 1032], 99.90th=[ 1144], 99.95th=[ 1448], 
| 99.99th=[ 2224] 

cpu server : 92.4 idle 
cpu client : 96,8 idle 




hardware (ceph node && client node): 
----------------------------------- 
ceph : hammer 
os : centos 7.1 
2 x 10cores Intel(R) Xeon(R) CPU E5-2687W v3 @ 3.10GHz 64GB ram 
2 x intel s3700 100GB : raid1: os + monitor 
6 x intel s3500 160GB : osds 
2x10gb mellanox connect-x3 (lacp) 

network 
------- 
mellanox sx1012 with breakout cables (10GB) 


centos tunning: 
--------------- 
-noop scheduler 
-tune-adm profile latency-performance 

ceph.conf 
--------- 
auth_cluster_required = cephx 
auth_service_required = cephx 
auth_client_required = cephx 
filestore_xattr_use_omap = true 


osd pool default min size = 1 

debug lockdep = 0/0 
debug context = 0/0 
debug crush = 0/0 
debug buffer = 0/0 
debug timer = 0/0 
debug journaler = 0/0 
debug osd = 0/0 
debug optracker = 0/0 
debug objclass = 0/0 
debug filestore = 0/0 
debug journal = 0/0 
debug ms = 0/0 
debug monc = 0/0 
debug tp = 0/0 
debug auth = 0/0 
debug finisher = 0/0 
debug heartbeatmap = 0/0 
debug perfcounter = 0/0 
debug asok = 0/0 
debug throttle = 0/0 

osd_op_threads = 5 
filestore_op_threads = 4 


osd_op_num_threads_per_shard = 1 
osd_op_num_shards = 10 
filestore_fd_cache_size = 64 
filestore_fd_cache_shards = 32 
ms_nocrc = true 
ms_dispatch_throttle_bytes = 0 

cephx sign messages = false 
cephx require signatures = false 

[client] 
rbd_cache = false 





rand 4K : rbd volume size: 10GB (data in osd node buffer - no access to disk) 
------------------------------------------------------------------------------ 
fio + librbd 
------------ 
[global] 
ioengine=rbd 
clientname=admin 
pool=pooltest 
rbdname=rbdtest 
invalidate=0 
rw=randread 
direct=1 
bs=4k 
numjobs=2 
group_reporting=1 
iodepth=32 



fio + krbd 
----------- 
[global] 
ioengine=aio 
invalidate=1 # mandatory 
rw=randread 
bs=4K 
direct=1 
numjobs=2 
group_reporting=1 
size=10G 

iodepth=32 
filename=/dev/rbd0 (noop scheduler) 






_______________________________________________ 
ceph-users mailing list 
ceph-users@xxxxxxxxxxxxxx 
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com





[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux