Hello, On Mon, 5 Feb 2018 22:04:00 +0100 Tobias Kropf wrote: > Hi ceph list, > > we have a hyperconvergent ceph cluster with kvm on 8 nodes with ceph > hammer 0.94.10. Do I smell Proxmox? > The cluster is now 3 years old an we plan with a new > cluster for a high iops project. We use replicated pools 3/2 and have > not the best latency on our switch backend. > > > ping -s 8192 10.10.10.40 > > 8200 bytes from 10.10.10.40: icmp_seq=1 ttl=64 time=0.153 ms > Not particularly great, yes. However your network latency is only one factor, Ceph OSDs add quite another layer there and do affect IOPS even more usually. For high IOPS you need of course fast storage, network AND CPUs. > > We plan to split the hyperconvergent setup to storage an compute nodes > and want to split ceph cluster and public network. Cluster network with > 40 gbit mellanox switches and public network with the existant 10gbit > switches. > You'd do a lot better if you were to go all 40Gb/s and forget about splitting networks. The faster replication network will: a) be underutilized all of the time in terms of bandwidth b) not help with read IOPS at all c) still be hobbled by the public network latency when it comes to write IOPS (but of course help in regards to replication latency). > Now my question... are 0.153ms - 0.170ms fast enough for the public > network? We must deploy a setup with 1500 - 2000 terminalserver.... > Define terminal server, are we talking Windows Virtual Desktops with RDP? Windows is quite the hog when it comes to I/O. Regards, Christian -- Christian Balzer Network/Systems Engineer chibi@xxxxxxx Rakuten Communications _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com