Hi all, ceph-ansible@stable-6.0 supports pacific and the current content in the branch 'master' (future stable-7.0) is intended to support Ceph Quincy. I can't speak on behalf of Dimitri but I'm personally willing to keep maintaining ceph-ansible if there are interests, but people must be aware that: - the official and supported installer will be cephadm, - officially, ceph-ansible will be unsupported with the consequences it might bring (testing efforts, ci ressources, etc.), - Fewer engineering efforts will be carried out (contributions are welcome!) Thanks, On Thu, 18 Mar 2021 at 11:26, Stefan Kooman <stefan@xxxxxx> wrote: > On 3/18/21 9:09 AM, Janne Johansson wrote: > > Den ons 17 mars 2021 kl 20:17 skrev Matthew H <matthew.heler@xxxxxxxxxxx > >: > >> > >> "A containerized environment just makes troubleshooting more difficult, > getting access and retrieving details on Ceph processes isn't as > straightforward as with a non containerized infrastructure. I am still not > convinced that containerizing everything brings any benefits except the > collocation of services." > >> > >> It changes the way you troubleshoot, but I don't find it more difficult > in the issues I have seen and had. Even today without containers, all > services can be co-located within the same hosts (mons,mgrs,osds,mds).. Is > there a situation you've seen where that has not been the case? > > > > New ceph users pop in all the time on the #ceph IRC and have > > absolutely no idea on how to see the relevant logs from the > > containerized services. > > While you might not need that much Ceph knowledge to get Ceph up and > running, it does require users to know how container deployments work. I > had to put quite a bit of work in to get what ceph-ansible was doing to > deploy the containers, and why it would fail (after some other tries). > You do need to have Ceph knowledge, still, when things do not go as > expected, and even beforehand to make the right decisions on how to set > up all the infrastructure. So arguably you need even more knowledge to > understand what is going on under the hood, be it Ceph or containers. > > > > > Me being one of the people that do run services on bare metal (and > > VMs) I actually can't help them, and it seems several other old ceph > > admins can't either. > > > > Not that it is impossible or might not even be hard to get them, but > > somewhere in the "it is so easy to get it up and running, just pop a > > container and off you go" docs there seem to be a lack of the parts > > "when the OSD crashes at boot, run this to export the file normally > > called /var/log/ceph/ceph-osd.12.log" meaning it becomes a black box > > to the users and they are left to wipe/reinstall or something else > > when it doesn't work. At the end, I guess the project will see less > > useful reports with Assert Failed logs from impossible conditions and > > more people turning away from something that could be fixed in the > > long run. > > There is a ceph manager module for that: > https://docs.ceph.com/en/latest/mgr/crash/ > > I guess an option to "always send crash logs to Ceph" could be build in. > If you trust Ceph with this data of course (opt-in). > > > > > > I get some of the advantages, and for stateless services elsewhere it > > might be gold to have containers, I am not equally enthusiastic about > > it for ceph. > > > > Yeah, so I think it's good to discuss pros and cons and see what problem > it solves, and what extra problems it creates. > > Gr. Stefan > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > -- *Guillaume Abrioux*Senior Software Engineer _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx