Re: slow perfomance: sanity check

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I've reduced OSDs to 12 and  moved journal to ssd drives and now have
"boost" with writes to ~33-35MB/s. Is it maximum without full ssd
pools?

Best,
Stan

2017-04-06 9:34 GMT+02:00 Stanislav Kopp <staskopp@xxxxxxxxx>:
> Hello,
>
> I'm evaluate ceph cluster, to see  if you can use it for our
> virtualization solution (proxmox). I'm using 3 nodes, running Ubuntu
> 16.04 with stock ceph (10.2.6), every OSD uses separate 8 TB spinning
> drive (XFS), MONITORs are installed on the same nodes, all nodes are
> connected via 10G switch.
>
> The problem is, on client I have only ~25-30 MB/s with seq. write. (dd
> with "oflag=direct"). Proxmox uses Firefly, which is old, I know.  But
> I have the same performance on my desktop running the same version as
> ceph nodes using rbd mount, iperf shows full speed (1GB or 10GB up to
> client).
> I know that this setup is not optimal and for production I will use
> separate MON nodes and ssd for OSDs, but was wondering is this
> performance still normal. This is my cluster status.
>
>      cluster 3ea55c7e-5829-46d0-b83a-92c6798bde55
>      health HEALTH_OK
>      monmap e5: 3 mons at
> {ceph01=10.1.8.31:6789/0,ceph02=10.1.8.32:6789/0,ceph03=10.1.8.33:6789/0}
>             election epoch 60, quorum 0,1,2 ceph01,ceph02,ceph03
>      osdmap e570: 42 osds: 42 up, 42 in
>             flags sortbitwise,require_jewel_osds
>       pgmap v14784: 1024 pgs, 1 pools, 23964 MB data, 6047 objects
>             74743 MB used, 305 TB / 305 TB avail
>                 1024 active+clean
>
> btw, bench on nodes itself looks good as far I see.
>
> ceph01:~# rados bench -p rbd 10 write
> ....
> Total time run:         10.159667
> Total writes made:      1018
> Write size:             4194304
> Object size:            4194304
> Bandwidth (MB/sec):     400.801
> Stddev Bandwidth:       38.2018
> Max bandwidth (MB/sec): 472
> Min bandwidth (MB/sec): 344
> Average IOPS:           100
> Stddev IOPS:            9
> Max IOPS:               118
> Min IOPS:               86
> Average Latency(s):     0.159395
> Stddev Latency(s):      0.110994
> Max latency(s):         1.1069
> Min latency(s):         0.0432668
>
>
> Thanks,
> Stan
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux