With filestore on XFS using SSD journals that have good O_DSYNC write
performance, we typically see between 60-80MB/s per disk before
replication for large object writes. This is assuming there are no
other bottlenecks or things going on though (pg splitting, recovery,
network issues, etc). Probably the best case scenario would be large
writes to an RBD volume with 4MB objects and enough PGs in the pool that
splits never need to happen.
Having said that, on setups where some of the drives are slow, the
network is misconfigured, there are too few PGs, there are too many
drives on one controller, or other issues, 25-30MB/s per disk is
certainly possible.
Mark
On 04/06/2017 10:05 AM, Stanislav Kopp wrote:
I've reduced OSDs to 12 and moved journal to ssd drives and now have
"boost" with writes to ~33-35MB/s. Is it maximum without full ssd
pools?
Best,
Stan
2017-04-06 9:34 GMT+02:00 Stanislav Kopp <staskopp@xxxxxxxxx>:
Hello,
I'm evaluate ceph cluster, to see if you can use it for our
virtualization solution (proxmox). I'm using 3 nodes, running Ubuntu
16.04 with stock ceph (10.2.6), every OSD uses separate 8 TB spinning
drive (XFS), MONITORs are installed on the same nodes, all nodes are
connected via 10G switch.
The problem is, on client I have only ~25-30 MB/s with seq. write. (dd
with "oflag=direct"). Proxmox uses Firefly, which is old, I know. But
I have the same performance on my desktop running the same version as
ceph nodes using rbd mount, iperf shows full speed (1GB or 10GB up to
client).
I know that this setup is not optimal and for production I will use
separate MON nodes and ssd for OSDs, but was wondering is this
performance still normal. This is my cluster status.
cluster 3ea55c7e-5829-46d0-b83a-92c6798bde55
health HEALTH_OK
monmap e5: 3 mons at
{ceph01=10.1.8.31:6789/0,ceph02=10.1.8.32:6789/0,ceph03=10.1.8.33:6789/0}
election epoch 60, quorum 0,1,2 ceph01,ceph02,ceph03
osdmap e570: 42 osds: 42 up, 42 in
flags sortbitwise,require_jewel_osds
pgmap v14784: 1024 pgs, 1 pools, 23964 MB data, 6047 objects
74743 MB used, 305 TB / 305 TB avail
1024 active+clean
btw, bench on nodes itself looks good as far I see.
ceph01:~# rados bench -p rbd 10 write
....
Total time run: 10.159667
Total writes made: 1018
Write size: 4194304
Object size: 4194304
Bandwidth (MB/sec): 400.801
Stddev Bandwidth: 38.2018
Max bandwidth (MB/sec): 472
Min bandwidth (MB/sec): 344
Average IOPS: 100
Stddev IOPS: 9
Max IOPS: 118
Min IOPS: 86
Average Latency(s): 0.159395
Stddev Latency(s): 0.110994
Max latency(s): 1.1069
Min latency(s): 0.0432668
Thanks,
Stan
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com