We are using a switchstack of Juniper EX4200 and EX3200. Cisco should work, too. Anotheroption is failover bonding (but multipathing with different ips is better) Mit freundlichen Gr??en / Best Regards, -- Consultant Dipl.-Inf. Uwe Grohnwaldt Gutleutstr. 351 60327 Frankfurt a. M. eMail: uwe at grohnwaldt.eu Telefon: +49-69-34878906 Mobil: +49-172-3209285 Fax: +49-69-348789069 ----- Original Message ----- > From: "Leen Besselink" <leen at consolejunkie.net> > To: ceph-users at lists.ceph.com > Cc: "Uwe Grohnwaldt" <uwe at grohnwaldt.eu> > Sent: Montag, 12. Mai 2014 19:01:46 > Subject: Re: Ceph with VMWare / XenServer > > On Mon, May 12, 2014 at 03:45:43PM +0200, Uwe Grohnwaldt wrote: > > Hi, > > > > yes, we use it in production. I can stop/kill the tgt on one server > > and XenServer goes to the second one. We enabled multipathing in > > xenserver. In our setup we haven't multiple ip-ranges so we > > scan/login the second target on xenserverstartup with iscsiadm in > > rc.local. > > > > Thats based on history - we used Dell Equallogic before ceph came > > in and there was no need to use multipathing (only LACP-channels). > > No we enabled multipathing and use tgt, but without diffent > > ip-ranges. > > > > I assume you connected the machines to the same switch ? As normal > LACP don't work with multiple switches. > > Is that correct ? > > It wasn't that I needed different ip-ranges in my setup, it just > makes it simpler/predictable. > > > Mit freundlichen Gr??en / Best Regards, > > -- > > Consultant > > Dipl.-Inf. Uwe Grohnwaldt > > Gutleutstr. 351 > > 60327 Frankfurt a. M. > > > > eMail: uwe at grohnwaldt.eu > > Telefon: +49-69-34878906 > > Mobil: +49-172-3209285 > > Fax: +49-69-348789069 > > > > ----- Original Message ----- > > > From: "Andrei Mikhailovsky" <andrei at arhont.com> > > > To: "Uwe Grohnwaldt" <uwe at grohnwaldt.eu> > > > Cc: ceph-users at lists.ceph.com > > > Sent: Montag, 12. Mai 2014 14:48:58 > > > Subject: Re: Ceph with VMWare / XenServer > > > > > > > > > Uwe, thanks for your quick reply. > > > > > > Do you run the Xenserver setup on production env and have you > > > tried > > > to test some failover scenarios to see if the xenserver guest vms > > > are working during the failover of storage servers? > > > > > > Also, how did you set up the xenserver iscsi? Have you used the > > > multipath option to set up the LUNs? > > > > > > Cheers > > > > > > > > > > > > > > > ----- Original Message ----- > > > > > > From: "Uwe Grohnwaldt" <uwe at grohnwaldt.eu> > > > To: ceph-users at lists.ceph.com > > > Sent: Monday, 12 May, 2014 12:57:48 PM > > > Subject: Re: Ceph with VMWare / XenServer > > > > > > Hi, > > > > > > at the moment we are using tgt with RBD backend compiled from > > > source > > > on Ubuntu 12.04 and 14.04 LTS. We have two machines within two > > > ip-ranges (e.g. 192.168.1.0/24 and 192.168.2.0/24). One machine > > > in > > > 192.168.1.0/24 and one machine in 192.168.2.0/24. The config for > > > tgt > > > is the same on both machines, they export the same rbd. This > > > works > > > well for XenServer. > > > > > > For VMWare you have to disable VAAI to use it with tgt > > > (http://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1033665) > > > If you don't disable it, ESXi becomes very slow and unresponsive. > > > > > > I think the problem is the iSCSI Write Same Support but I haven't > > > tried which of the settings of VAAI is responsible for this > > > behavior. > > > > > > Mit freundlichen Gr??en / Best Regards, > > > -- > > > Consultant > > > Dipl.-Inf. Uwe Grohnwaldt > > > Gutleutstr. 351 > > > 60327 Frankfurt a. M. > > > > > > eMail: uwe at grohnwaldt.eu > > > Telefon: +49-69-34878906 > > > Mobil: +49-172-3209285 > > > Fax: +49-69-348789069 > > > > > > ----- Original Message ----- > > > > From: "Andrei Mikhailovsky" <andrei at arhont.com> > > > > To: ceph-users at lists.ceph.com > > > > Sent: Montag, 12. Mai 2014 12:00:48 > > > > Subject: Ceph with VMWare / XenServer > > > > > > > > > > > > > > > > Hello guys, > > > > > > > > I am currently running a ceph cluster for running vms with qemu > > > > + > > > > rbd. It works pretty well and provides a good degree of > > > > failover. I > > > > am able to run maintenance tasks on the ceph nodes without > > > > interrupting vms IO. > > > > > > > > I would like to do the same with VMWare / XenServer > > > > hypervisors, > > > > but > > > > I am not really sure how to achieve this. Initially I thought > > > > of > > > > using iscsi multipathing, however, as it turns out, > > > > multipathing is > > > > more for load balancing and nic/switch failure. It does not > > > > allow > > > > me > > > > to perform maintenance on the iscsi target without interrupting > > > > service to vms. > > > > > > > > Has anyone done either a PoC or better a production environment > > > > where > > > > they've used ceph as a backend storage with vmware / xenserver? > > > > The > > > > important element for me is to have the ability of performing > > > > maintenance tasks and resilience to failovers without > > > > interrupting > > > > IO to vms. Are there any recommendations or howtos on how this > > > > could > > > > be achieved? > > > > > > > > Many thanks > > > > > > > > Andrei > > > > > > > > > > > > _______________________________________________ > > > > ceph-users mailing list > > > > ceph-users at lists.ceph.com > > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > > > _______________________________________________ > > > ceph-users mailing list > > > ceph-users at lists.ceph.com > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > > > > _______________________________________________ > > ceph-users mailing list > > ceph-users at lists.ceph.com > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >