Re: cephadm docs on HA NFS

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



The design is for cephadm to resurrect the failed ganesha rank on
another host, just like k8s does.  The current gap in the
implementation is that the cephadm refresh is pretty slow (~10m) which
means that probably won't happen before the NFS state times out.  We
are working on improving the refresh/responsiveness right now, but
you're right that the current code isn't yet complete.

I think a PR that updates the docs to note that the ingress mode isn't
yet complete, and also adds a section that documents how to do
round-robin DNS would be the best immediate step?

sage


On Thu, Jul 22, 2021 at 9:32 AM Jeff Layton <jlayton@xxxxxxxxxx> wrote:
>
> I think we probably need to redo this bit of documentation:
>
>     https://docs.ceph.com/en/latest/cephadm/nfs/#high-availability-nfs
>
> I would just spin up a patch, but I think we might also just want to
> reconsider recommending an ingress controller at all.
>
> Some people seem to be taking this to mean that they can shoot down one
> of the nodes in the NFS server cluster, and the rest will just pick up
> the load. That's not at all how this works.
>
> If a NFS cluster node goes down, then it _must_ be resurrected in some
> fashion, period. Otherwise, the MDS will eventually (in 5 mins) time out
> the state it held and the NFS clients will not be able to reclaim their
> state.
>
> Given that, the bulleted list at the start of the doc above is wrong. We
> cannot do any sort of failover if there is a host failure. My assumption
> was that the orchestrator took care of starting up an NFS server
> elsewhere if the host it was running on went down. Is that not the case?
>
> In any case, think we should reconsider recommending an ingress
> controller at all. It's really just another point of failure, and a lot
> of people seem to be misconstruing what guarantees that offers.
>
> Round-robin DNS would be a better option in this situation, and it
> wouldn't be as problematic if we want to support things like live
> shrinking the cluster in the future.
> --
> Jeff Layton <jlayton@xxxxxxxxxx>
>
> _______________________________________________
> Dev mailing list -- dev@xxxxxxx
> To unsubscribe send an email to dev-leave@xxxxxxx
_______________________________________________
Dev mailing list -- dev@xxxxxxx
To unsubscribe send an email to dev-leave@xxxxxxx



[Index of Archives]     [CEPH Users]     [Ceph Devel]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux