On 01/29/2018 06:33 PM, Nico Schottelius wrote:
Good evening list, we are soon expanding our data center [0] to a new location [1]. We are mainly offering VPS / VM Hosting, so rbd is our main interest. We have a low latency 10 Gbit/s link between our other location [2] and we are wondering, what is the best practise for expanding.
What is 'low latency'? If you are not using RBD mirroring and are trying to span a Ceph cluster over two DCs you will usually run into latency problems.
Any increase in latency will lower the IOps and decrease performance.
Naturally we think about creating a new ceph cluster that is independent from the first location, so connection interrupts (unlikely) or different power outages (more likely) are becoming a concern. Given that we running two different ceph clusters, we think about rbd mirroring, so that we can (partially) mirror one side to the other or vice versa. However using this approach we lose the possibility to have very big rbd images (big as in 10ths to 100ds of TBs), as the storage is divided. My question to the list is, how have you handled this situation so far? Would you also recommend splitting or have you expanded ceph clusters over several kilometers of range so far? With what experiences?
Like I said, latency, latency, latency. That's what matters. Bandwidth usually isn't a real problem.
What latency do you have with a 8k ping between hosts? Wido
I am very curious to hear your answers! Best, Nico [0] https://datacenterlight.ch [1] Linthal, in pretty Glarus https://www.google.ch/maps/place/Linthal,+8783+Glarus+S%C3%BCd/ [2] Schwanden, also pretty https://www.google.ch/maps/place/Schwanden,+8762+Glarus+S%C3%BCd/ -- Modern, affordable, Swiss Virtual Machines. Visit www.datacenterlight.ch _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com