The design is for cephadm to resurrect the failed ganesha rank on another host, just like k8s does. The current gap in the implementation is that the cephadm refresh is pretty slow (~10m) which means that probably won't happen before the NFS state times out. We are working on improving the refresh/responsiveness right now, but you're right that the current code isn't yet complete. I think a PR that updates the docs to note that the ingress mode isn't yet complete, and also adds a section that documents how to do round-robin DNS would be the best immediate step? sage On Thu, Jul 22, 2021 at 9:32 AM Jeff Layton <jlayton@xxxxxxxxxx> wrote: > > I think we probably need to redo this bit of documentation: > > https://docs.ceph.com/en/latest/cephadm/nfs/#high-availability-nfs > > I would just spin up a patch, but I think we might also just want to > reconsider recommending an ingress controller at all. > > Some people seem to be taking this to mean that they can shoot down one > of the nodes in the NFS server cluster, and the rest will just pick up > the load. That's not at all how this works. > > If a NFS cluster node goes down, then it _must_ be resurrected in some > fashion, period. Otherwise, the MDS will eventually (in 5 mins) time out > the state it held and the NFS clients will not be able to reclaim their > state. > > Given that, the bulleted list at the start of the doc above is wrong. We > cannot do any sort of failover if there is a host failure. My assumption > was that the orchestrator took care of starting up an NFS server > elsewhere if the host it was running on went down. Is that not the case? > > In any case, think we should reconsider recommending an ingress > controller at all. It's really just another point of failure, and a lot > of people seem to be misconstruing what guarantees that offers. > > Round-robin DNS would be a better option in this situation, and it > wouldn't be as problematic if we want to support things like live > shrinking the cluster in the future. > -- > Jeff Layton <jlayton@xxxxxxxxxx> > > _______________________________________________ > Dev mailing list -- dev@xxxxxxx > To unsubscribe send an email to dev-leave@xxxxxxx _______________________________________________ Dev mailing list -- dev@xxxxxxx To unsubscribe send an email to dev-leave@xxxxxxx