Re: autoconfigured haproxy service?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Jul 11, 2017 at 5:40 PM, Sage Weil <sage@xxxxxxxxxxxx> wrote:
> On Tue, 11 Jul 2017, Haomai Wang wrote:
>> On Tue, Jul 11, 2017 at 11:11 PM, Sage Weil <sage@xxxxxxxxxxxx> wrote:
>> > On Tue, 11 Jul 2017, Sage Weil wrote:
>> >> Hi all,
>> >>
>> >> Luminous features a new 'service map' that lets rgw's (and rgw nfs
>> >> gateways and iscsi gateways and rbd mirror daemons and ...) advertise
>> >> themselves to the cluster along with some metadata (like the addresses
>> >> they are binding to and the services the provide).
>> >>
>> >> It should be pretty straightforward to build a service that
>> >> auto-configures haproxy based on this information so that you can deploy
>> >> an rgw front-end that dynamically reconfigures itself when additional
>> >> rgw's are deployed or removed.  haproxy has a facility to adjust its
>> >> backend configuration at runtime[1].
>> >>
>> >> Anybody interested in tackling this?  Setting up the load balancer in
>> >> front of rgw is one of the more annoying pieces of getting ceph up and
>> >> running in production and until now has been mostly treated as out of
>> >> scope.  It would be awesome if there was an autoconfigured service that
>> >> did it out of the box (and had all the right haproxy options set).
>> >
>> > [1] https://stackoverflow.com/questions/42678269/haproxy-dynamic-configuration
>>
>> it looks we do more compared to before. do we need to care the
>> lifecycle of haproxy?  we need to manage haproxy in ceph command?
>
> I don't think so, although not having done this much I'm not the
> expert.
>
> My suggestion would be a new package like radosgw-haproxy-agent that
> depends on haproxy and includes a script and some systemd units etc so
> that with minimal configuration (i.e., set up ceph.conf auth key or
> something) it will wake up periodically and refresh the running haproxy's
> config.

So IIUC you want to periodically discover the set of radosgw backends
to fill haproxy.cfg, then reload the haproxy daemons. That would be
useful to (a) keep the set of radosgw hosts up to date and (b) to
provide a high quality haproxy configuration OOTB.

The stackoverflow link you sent is about another interesting use-case
of haproxy -- mapping different urls to different backends. Indeed we
used this in the past to migrate between ceph clusters, bucket by
bucket. And we still use it today to redirect a few very busy buckets
to an isolated set of radosgw's. I can share our config if that helps
explain how this works [2]. And maybe that config can already start a
debate about which are the best settings for an haproxy frontend (I
won't claim ours is generally correct -- happy to hear about how it
could be improved).

I don't know if the bucket mapping concept is generally applicable.
Maybe this haproxy-agent should focus on configuring a single backend
populated with the radosgw's, and leave more complex configurations up
to their admins?

(BTW, we generate this haproxy.cfg dynamically via puppet, which fills
a template by discovering the radosgw hosts in our PuppetDB).

Cheers, Dan

[2] https://gist.github.com/dvanders/857ffcf7249849cffc8d784c55b1a4d5

> We could add a 'ceph-deploy haproxy create ...' command to deploy it,
> along with something similar in ceph-ansible...
>
> sage
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux