Re: monitoring

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 2019-12-04T00:12:27, Sage Weil <sage@xxxxxxxxxxxx> wrote:

> > Perhaps, all of these design points trace back to a single idea - support
> > multiple ceph clusters on the same set of machine(s). Is this the goal? Is
> > this want Ceph users want?
> It was one of my goals.  A few reasons:

So I agree this is useful. Glad to have this on the map again. I'm not
sure how large this percentage is - the environments that I am aware of
that'd require segregation into multiple clusters would also frown on
using the same CPU/OS instance for running them, split out disks
wouldn't be enough -, but I see it could be helpful.


On the soapbox, though:

> > Now picking up on the scope issue for the orchestrator - apologies if this
> > sounds like a manifesto...I'm a "usability" addict!
> > 
> > IMO, our collective goal should be to drive ease of use and Ceph adoption
> > beyond Linux geeks. If that's a view that resonates, I think the
> > orchestrator has critical role to play to enable that strategy
> > Personally I'd would like to see the orchestrator evolve over time to
> > become the automation engine that enables an open source ecosystem around
> > Ceph;
> > - provide a default implementation for monitoring/alerting/metrics - this
> > can be simple and doesn't need HA - as Sage has already mentioned

Doesn't it though - not being alerted when your cluster enters a degraded
state is pretty dangerous from an operational point of view. And as soon
as users start relying on those metrics, losing them is also not an
option.

We can't not deploy monitoring/alerting/metrics, since they're essential
to operating a Ceph cluster. But I think we're deluding ourselves if we
even pretend it'll remain that simple - because then design decisions
will not consider the endstate.

> > - samba/ganesha deployment, loud balancers to improve radosgw etc etc

Managing the access protocols/gateways is something we'll have to do.
And again, that means we'll have to take HA into account.

I'm a wee bit on the fence on the LB parts. We surely need to have hooks
to inform the LB layer about which endpoints we deployed, but
managing/deploying/configuring the LBs themselves I'd hope is out of
scope.

> > - integration with platform management (why not show in the ceph dashboard
> > whether you have patches outstanding against your host, or the host has a
> > bad PSU) - enable the sysadmin to work more efficiently on Ceph, and maybe
> > they'd prefer it over other platforms.

Why not? Because those are system management tasks that are outside
Ceph.

Please don't build an inferior salt/ansible/puppet/chef or systems
management console. Solutions for these problems exist. Let's not
duplicate everything.

> > </soapbox>
> +1

-1



Regards,
    Lars

-- 
SUSE Software Solutions Germany GmbH, MD: Felix Imendörffer, HRB 36809 (AG Nürnberg)
"Architects should open possibilities and not determine everything." (Ueli Zbinden)
_______________________________________________
Dev mailing list -- dev@xxxxxxx
To unsubscribe send an email to dev-leave@xxxxxxx




[Index of Archives]     [CEPH Users]     [Ceph Devel]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux