Hi, I’m coming at this from the position of a newbie to Ceph. I had some experience of it as part of Proxmox, but not as a standalone solution. I really don’t care whether Ceph is contained or not, I don’t have the depth of knowledge or experience to argue it either way. I can see that containers may well offer a more consistent deployment scenario with fewer dependencies on the external host OS. Upgrades/patches to the host OS may not impact the container deployment etc., with the two systems not held in any lock-step. The challenge for me hasn’t been Ceph its self. Ceph has worked brilliantly, I have a fully resilient architecture split between two active datacentres and my storage can survive up-to 50% node/OSD hardware failure. No, the challenge has been documentation. I’ve run off down multiple rabbit holes trying to find solutions to problems or just background information. I’ve been tripped up by not spotting the Ceph documentation was “v: latest” rather than “v: octopus”...so features didn’t exist or commands were structured slightly differently. Also just not being obvious whether the bit of documentation I was looking at related to a native Ceph package deployment or a container one. Plus you get the Ceph/Suse/Redhat/Proxmox/IBM etc..etc.. flavour answer depending on which Google link you click. Yes I know, its part of the joy of working with open source....but still, not what you need when I chunk of infrastructure has failed and you don’t know why. I’m truly in awe of what the Ceph community has produced and is planning for the future, so don’t think I’m any kind of hater. My biggest request is for the documentation to take on some restructuring. Keep the different deployment methods documented separately, yes an intro covering the various options and recommendations is great, but then keep it entirely discreet. Then when a feature/function is documented, make it clear if this applies to packaged or container deployment etc... e.g. Zabbix (we use Zabbix)....lovely documentation on how to integrate Ceph and Zabbix....until you finally find out its not supported with containers....via a forum and an RFE/Bug entry. And thank you to all the support in the community, REALLY appreciated. Best, Andrew Sent from Mail<https://go.microsoft.com/fwlink/?LinkId=550986> for Windows From: Erik Lindahl<mailto:erik.lindahl@xxxxxxxxx> Sent: 17 August 2021 16:01 To: Marc<mailto:Marc@xxxxxxxxxxxxxxxxx> Cc: Nico Schottelius<mailto:nico.schottelius@xxxxxxxxxxx>; Kai Börnert<mailto:kai.boernert@xxxxxxxxx>; ceph-users<mailto:ceph-users@xxxxxxx> Subject: Re: Why you might want packages not containers for Ceph deployments Hi, Whether containers are good or not is a separate discussion where I suspect there won't be consensus in the near future. However, after just having looked at the documentation again, my main point would be that when a major stable open source project recommends a specific installation method (=cephadm) first in the "getting started" guide, users are going to expect that's the alternative things are documented for, which isn't quite the case for cephadm (yet). Most users will probably accept either solution as long as there is ONE clear & well-documented way of working with ceph - but the current setup of even having the simple (?) getting started guide talk about at least three different ways without clearly separating their documentation seems like a guarantee for long-term confusion and higher entry barriers for new users, which I assume is the opposite of the goal of cephadm! Cheers, Erik Erik Lindahl <erik.lindahl@xxxxxxxxxxxxx> Professor of Biophysics Science for Life Laboratory Stockholm University & KTH Office (SciLifeLab): +46 8 524 81567 Cell (Sweden): +46 73 4618050 Cell (US): +1 (650) 924 7674 > On 17 Aug 2021, at 16:29, Marc <Marc@xxxxxxxxxxxxxxxxx> wrote: > > >> >> >> Again, this is meant as hopefully constructive feedback rather than >> complaints, but the feeling a get after having had fairly smooth >> operations with raw packages (including fixing previous bugs leading to >> severe crashes) and lately grinding our teeth a bit over cephadm is that >> it has helped automated a bunch of stuff that wasn't particularly >> difficult (it's nice to issue an update with a single command, but it >> works perfectly fine manually too) at the cost of making it WAY more >> difficult to fix things (not to mention simply get information about the >> cluster) when we have problems - and in the long run that's not a trade- >> off I'm entirely happy with :-) >> > > Everyone can only agree to keeping things simple. I honestly do not even know why you want to try cephadm. The containerized solution was developed to replace ceph deploy, ceph ansible etc. as a solution to make ceph installation for new users easier. That is exactly the reason (imho) why you should not use the containerized environment. Because a containerized environment has not as primaray task being an easy deployment tool. And because the focus is on easy deployment, the real characteristics of the containerized environment are being ignored during this development. Such as, you must be out of your mind to create a depency between ceph-osd/msd/mon/all and dockerd. > > 10 years(?) ago the people of mesos thought the docker containerizer was 'flacky' and created their own more stable containerizer. And still today, containers are being killed if dockerd is terminated. What some users had to learn the hard way, as recently posted here. > > Today's container solutions are not on the level where you can say, you require absolutely no knowledge to fix issues. So that means you would always require knowledge of the container solution + ceph to troubleshoot. And that is of course more knowledge, than just knowing ceph. > > I would not be surprised if cephadm ends up like ceph deploy/ansible. _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx