Re: OSPF to the host

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




> -----Original Message-----
> From: Luis Periquito [mailto:periquito@xxxxxxxxx]
> Sent: 06 June 2016 14:30
> To: Nick Fisk <nick@xxxxxxxxxx>
> Cc: Ceph Users <ceph-users@xxxxxxxxxxxxxx>
> Subject: Re:  OSPF to the host
> 
> Nick,
> 
> TL;DR: works brilliantly :)

Excellent, just what I wanted to hear!!!

> 
> Where I work we have all of the ceph nodes (and a lot of other stuff) using
> OSPF and BGP server attachment. With that we're able to implement
> solutions like Anycast addresses, removing the need to add load balancers,
> for the radosgw solution.
> 
> The biggest issues we've had were around the per-flow vs per-packets traffic
> load balancing, but as long as you keep it simple you shouldn't have any
> issues.
> 
> Currently we have a P2P network between the servers and the ToR switches
> on a /31 subnet, and then create a virtual loopback address, which is the
> interface we use for all communications. Running tests like iperf we're able
> to reach 19Gbps (on a 2x10Gbps network). OTOH we no longer have the
> ability to separate traffic between public and osd network, but never really
> felt the need for it.

Yeah, I've come to pretty much the same conclusion.

> 
> Also spend a bit of time planning how the network will look like and it's
> topology. If done properly (think details like route summarization) then it's
> really worth the extra effort.

How are you doing the route summarization with OSPF? Is each rack (for example) a separate OSPF area, which you then summarize and send up to area 0?

> 
> 
> 
> On Mon, Jun 6, 2016 at 11:57 AM, Nick Fisk <mailto:nick@xxxxxxxxxx> wrote:
> Hi All,
> 
> Has anybody had any experience with running the network routed down all
> the way to the host?
> 
> I know the standard way most people configured their OSD nodes is to bond
> the two nics which will then talk via a VRRP gateway and then probably from
> then on the networking is all Layer3. The main disadvantage I see here is that
> you need a beefy inter switch link to cope with the amount of traffic flowing
> between switches to the VRRP address. I’ve been trying to design around
> this by splitting hosts into groups with different VRRP gateways on either
> switch, but this relies on using active/passive bonding on the OSD hosts to
> make sure traffic goes from the correct Nic to the directly connected switch.
> 
> What I was thinking, instead of terminating the Layer3 part of the network at
> the access switches, terminate it at the hosts. If each Nic of the OSD host had
> a different subnet and the actual “OSD Server” address bound to a loopback
> adapter, OSPF should advertise this loopback adapter address as reachable
> via the two L3 links on the physically attached Nic’s. This should give you a
> redundant topology which also will respect your physically layout and
> potentially give you higher performance due to ECMP.
> 
> Any thoughts, any pitfalls?
> 
> 
> 
> _______________________________________________
> ceph-users mailing list
> mailto:ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux