slow perfomance: sanity check

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello,

I'm evaluate ceph cluster, to see  if you can use it for our
virtualization solution (proxmox). I'm using 3 nodes, running Ubuntu
16.04 with stock ceph (10.2.6), every OSD uses separate 8 TB spinning
drive (XFS), MONITORs are installed on the same nodes, all nodes are
connected via 10G switch.

The problem is, on client I have only ~25-30 MB/s with seq. write. (dd
with "oflag=direct"). Proxmox uses Firefly, which is old, I know.  But
I have the same performance on my desktop running the same version as
ceph nodes using rbd mount, iperf shows full speed (1GB or 10GB up to
client).
I know that this setup is not optimal and for production I will use
separate MON nodes and ssd for OSDs, but was wondering is this
performance still normal. This is my cluster status.

     cluster 3ea55c7e-5829-46d0-b83a-92c6798bde55
     health HEALTH_OK
     monmap e5: 3 mons at
{ceph01=10.1.8.31:6789/0,ceph02=10.1.8.32:6789/0,ceph03=10.1.8.33:6789/0}
            election epoch 60, quorum 0,1,2 ceph01,ceph02,ceph03
     osdmap e570: 42 osds: 42 up, 42 in
            flags sortbitwise,require_jewel_osds
      pgmap v14784: 1024 pgs, 1 pools, 23964 MB data, 6047 objects
            74743 MB used, 305 TB / 305 TB avail
                1024 active+clean

btw, bench on nodes itself looks good as far I see.

ceph01:~# rados bench -p rbd 10 write
....
Total time run:         10.159667
Total writes made:      1018
Write size:             4194304
Object size:            4194304
Bandwidth (MB/sec):     400.801
Stddev Bandwidth:       38.2018
Max bandwidth (MB/sec): 472
Min bandwidth (MB/sec): 344
Average IOPS:           100
Stddev IOPS:            9
Max IOPS:               118
Min IOPS:               86
Average Latency(s):     0.159395
Stddev Latency(s):      0.110994
Max latency(s):         1.1069
Min latency(s):         0.0432668


Thanks,
Stan
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux