We have a fairly significant number of customers running our Ceph product (SUSE Enterprise Storage) which leverages horizontally scaling LIO based gateways. There has been kernel work
done to deal with some VMware specific requirements of the protocol. Overall, performance is good. There are a few different approaches to mitigate the timeouts that happen for OSD failure detection.
1 – tune the thresholds for failure detection 2 – use hardware raid The first is the more desirable as it maintains purity of the Ceph environment. I’ll second the plug for using KVM instead of VMware. A hypervisor with a native protocol connector will outperform one going through a gateway any day. David Byte Sr. Technology Strategist Alliances and SUSE Embedded dbyte@xxxxxxxx 918.528.4422 From:
ceph-users <ceph-users-bounces@xxxxxxxxxxxxxx> on behalf of Maged Mokhtar <mmokhtar@xxxxxxxxxxx> Hi, Please check the PetaSAN project
We provide clustered iSCSI using LIO/Ceph rbd and Consul for HA. Works well with VMWare. /Maged From:
Osama Hasebou
Sent: Thursday, June 15, 2017 12:29 PM To:
ceph-users
Subject: [ceph-users] VMware + CEPH Integration Hi Everyone, We would like to start testing using VMware with CEPH storage. Can people share their experience with production ready ideas they tried and if they were successful? I have been reading lately that either NFS or iSCSI are possible with some server acting as a gateway in between Ceph and VMware environment but NFS is better. Thank you. Regards, _______________________________________________ |
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com