Re: full ssd setup preliminary hammer bench

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




Am 17.04.2015 um 17:37 schrieb Alexandre DERUMIER:
Hi Mark,

I finally got my hardware for my production full ssd cluster.

Here a first preliminary bench. (1osd).

I got around 45K iops with randread 4K with a small 10GB rbd volume


I'm pretty happy because I don't see anymore huge cpu difference between krbd && lirbd.
In my previous bench I was using debian wheezy as client,
now it's a centos 7.1, so maybe something is different (glibc,...).

any idea whether this might be the tcmalloc bug?


I'm planning to do big benchmark centos vs ubuntu vs debian, client && server, to compare.
I have 18 osd ssd for the benchmarks.







results : rand 4K : 1 osd
-------------------------

fio + librbd:
------------
iops: 45.1K

     clat percentiles (usec):
      |  1.00th=[  358],  5.00th=[  406], 10.00th=[  446], 20.00th=[  556],
      | 30.00th=[  676], 40.00th=[ 1048], 50.00th=[ 1192], 60.00th=[ 1304],
      | 70.00th=[ 1400], 80.00th=[ 1496], 90.00th=[ 1624], 95.00th=[ 1720],
      | 99.00th=[ 1880], 99.50th=[ 1928], 99.90th=[ 2064], 99.95th=[ 2128],
      | 99.99th=[ 2512]

cpu server :  89.1 iddle
cpu client :  92,5 idle

fio + krbd
----------
iops:47.5K

     clat percentiles (usec):
      |  1.00th=[  620],  5.00th=[  636], 10.00th=[  644], 20.00th=[  652],
      | 30.00th=[  668], 40.00th=[  676], 50.00th=[  684], 60.00th=[  692],
      | 70.00th=[  708], 80.00th=[  724], 90.00th=[  756], 95.00th=[  820],
      | 99.00th=[ 1004], 99.50th=[ 1032], 99.90th=[ 1144], 99.95th=[ 1448],
      | 99.99th=[ 2224]

cpu server :  92.4 idle
cpu client :  96,8 idle




hardware (ceph node && client node):
-----------------------------------
ceph : hammer
os : centos 7.1
2 x 10cores Intel(R) Xeon(R) CPU E5-2687W v3 @ 3.10GHz
64GB ram
2 x intel s3700 100GB : raid1: os + monitor
6 x intel s3500 160GB : osds
2x10gb mellanox connect-x3 (lacp)

network
-------
mellanox sx1012 with breakout cables (10GB)


centos tunning:
---------------
-noop scheduler
-tune-adm profile latency-performance

ceph.conf
---------
auth_cluster_required = cephx
auth_service_required = cephx
auth_client_required = cephx
filestore_xattr_use_omap = true


         osd pool default min size = 1

         debug lockdep = 0/0
         debug context = 0/0
         debug crush = 0/0
         debug buffer = 0/0
         debug timer = 0/0
         debug journaler = 0/0
         debug osd = 0/0
         debug optracker = 0/0
         debug objclass = 0/0
         debug filestore = 0/0
         debug journal = 0/0
         debug ms = 0/0
         debug monc = 0/0
         debug tp = 0/0
         debug auth = 0/0
         debug finisher = 0/0
         debug heartbeatmap = 0/0
         debug perfcounter = 0/0
         debug asok = 0/0
         debug throttle = 0/0

         osd_op_threads = 5
         filestore_op_threads = 4


         osd_op_num_threads_per_shard = 1
         osd_op_num_shards = 10
         filestore_fd_cache_size = 64
         filestore_fd_cache_shards = 32
         ms_nocrc = true
         ms_dispatch_throttle_bytes = 0

         cephx sign messages = false
         cephx require signatures = false

[client]
rbd_cache = false





rand 4K : rbd volume size: 10GB  (data in osd node buffer - no access to disk)
------------------------------------------------------------------------------
fio + librbd
------------
[global]
ioengine=rbd
clientname=admin
pool=pooltest
rbdname=rbdtest
invalidate=0
rw=randread
direct=1
bs=4k
numjobs=2
group_reporting=1
iodepth=32



fio + krbd
-----------
[global]
ioengine=aio
invalidate=1    # mandatory
rw=randread
bs=4K
direct=1
numjobs=2
group_reporting=1
size=10G

iodepth=32
filename=/dev/rbd0   (noop scheduler)






_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux