Hello, I'm evaluate ceph cluster, to see if you can use it for our virtualization solution (proxmox). I'm using 3 nodes, running Ubuntu 16.04 with stock ceph (10.2.6), every OSD uses separate 8 TB spinning drive (XFS), MONITORs are installed on the same nodes, all nodes are connected via 10G switch. The problem is, on client I have only ~25-30 MB/s with seq. write. (dd with "oflag=direct"). Proxmox uses Firefly, which is old, I know. But I have the same performance on my desktop running the same version as ceph nodes using rbd mount, iperf shows full speed (1GB or 10GB up to client). I know that this setup is not optimal and for production I will use separate MON nodes and ssd for OSDs, but was wondering is this performance still normal. This is my cluster status. cluster 3ea55c7e-5829-46d0-b83a-92c6798bde55 health HEALTH_OK monmap e5: 3 mons at {ceph01=10.1.8.31:6789/0,ceph02=10.1.8.32:6789/0,ceph03=10.1.8.33:6789/0} election epoch 60, quorum 0,1,2 ceph01,ceph02,ceph03 osdmap e570: 42 osds: 42 up, 42 in flags sortbitwise,require_jewel_osds pgmap v14784: 1024 pgs, 1 pools, 23964 MB data, 6047 objects 74743 MB used, 305 TB / 305 TB avail 1024 active+clean btw, bench on nodes itself looks good as far I see. ceph01:~# rados bench -p rbd 10 write .... Total time run: 10.159667 Total writes made: 1018 Write size: 4194304 Object size: 4194304 Bandwidth (MB/sec): 400.801 Stddev Bandwidth: 38.2018 Max bandwidth (MB/sec): 472 Min bandwidth (MB/sec): 344 Average IOPS: 100 Stddev IOPS: 9 Max IOPS: 118 Min IOPS: 86 Average Latency(s): 0.159395 Stddev Latency(s): 0.110994 Max latency(s): 1.1069 Min latency(s): 0.0432668 Thanks, Stan _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com