Setting 'osd addr' in the osd configuration section unfortunately also does not influence source address selection, the outgoing interface IP is used like before. On Thu, 7 Jul 2016, Luis Periquito wrote: > If, like me, you have several different networks, or they overlap for > whatever reason, I just have the options: > > mon addr = IP:port > osd addr = IP > > in the relevant sections. However I use puppet to deploy ceph, and all > files are "manually" created. > > So it becomes something like this: > > [mon.mon1] > host = mon1 > mon addr = x.y.z.a:6789 > mon data = /var/lib/ceph/mon/internal-mon1 > [osd.0] > host = dskh1 > osd addr = x.y.z.a > osd data = /var/lib/ceph/osd/osd-0 > osd journal = /var/lib/ceph/osd/journal/osd-0 > keyring = /var/lib/ceph/osd/osd-0/keyring > osd max backfills = 1 > osd recovery max active = 1 > osd recovery op priority = 1 > osd client op priority = 63 > osd disk thread ioprio class = idle > osd disk thread ioprio priority = 7 > [osd.1] > .... > > necessarily the host and addr parts are correct in our environment. > > On 7 July 2016 at 11:36, Nick Fisk <nick@xxxxxxxxxx> wrote: > > > > -----Original Message----- > > > From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf > > Of Matyas Koszik > > > Sent: 07 July 2016 11:26 > > > To: ceph-users@xxxxxxxxxxxxxx > > > Subject: layer3 network > > > > > > > > > > > > Hi, > > > > > > My setup uses a layer3 network, where each node has two connections > > (/31s), equipped with a loopback address and redundancy is > > > provided via OSPF. In this setup it is important to use the loopback > > address as source for outgoing connections, since the > > interface > > > addresses are not protected from failure, but the loopback address is. > > > > > > So I set the public addr and the cluster addr to the desired ip, but it > > seems that the outgoing connections do not use this as the > > source > > > address. > > > I'm using jewel; is this the expected behavior? > > > > Do your public/cluster networks overlap the physical connection IP's? From > > what I understand Ceph binds to the interface whose IP > > lies within the range specified in the conf file. > > > > So for example if public addr = 192.168.1.0/24 > > > > Then your loopback should be in that range, but you must make sure the > > physical nics lie outside this range. > > > > I'm following this with interest as I am about to deploy something very > > similar. > > > > > > > > Matyas > > > > > > > > > _______________________________________________ > > > ceph-users mailing list > > > ceph-users@xxxxxxxxxxxxxx > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > _______________________________________________ > > ceph-users mailing list > > ceph-users@xxxxxxxxxxxxxx > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > > -- > > Luis Periquito > > Unix Team Lead > > <http://www.ocado.com/> > > Head Office, Titan Court, 3 Bishop Square, Hatfield Business Park, > Hatfield, Herts AL10 9NE > > -- > > > Notice: This email is confidential and may contain copyright material of > members of the Ocado Group. Opinions and views expressed in this message > may not necessarily reflect the opinions and views of the members of the > Ocado Group. > > > > If you are not the intended recipient, please notify us immediately and > delete all copies of this message. Please note that it is your > responsibility to scan this message for viruses. > > > > Fetch and Sizzle are trading names of Speciality Stores Limited, a member > of the Ocado Group. > > > > References to the â??Ocado Groupâ?? are to Ocado Group plc (registered in > England and Wales with number 7098618) and its subsidiary undertakings (as > that expression is defined in the Companies Act 2006) from time to time. > The registered office of Ocado Group plc is Titan Court, 3 Bishops Square, > Hatfield Business Park, Hatfield, Herts. AL10 9NE. > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com