On Thu, May 31, 2018 at 10:33 PM, Marc Roos <M.Roos@xxxxxxxxxxxxxxxxx> wrote: > > I actually tried to search the ML before bringing up this topic. Because > I do not get the logic choosing this direction. > > - Bluestore is created to cut out some fs overhead, > - everywhere 10Gb is recommended because of better latency. (I even > posted here something to make ceph better performing with 1Gb eth, > disregarded because it would add complexity, fine, I can understand) > > And then because of some start-up/automation issues, lets add the lvm > tier? Introducing a layer that is constantly there and adds some > overhead (maybe not that much) for every read and write operation? > > Is see ceph-disk as a tool to prepare the osd and the do the rest > myself. Without ceph-deploy or ansible, because I trust more what I see > I type than someone else scripted. I don’t have any startup problems. You can certainly do that with ceph-volume. You can create the OSD manually, and then add the information about your OSD (drives, locations, fsid, uuids, etc..) on /etc/ceph/osd/ This is how we are able to take over ceph-disk deployed OSDs See: http://docs.ceph.com/docs/master/ceph-volume/simple/scan/#scan > > Do assume I am not an expert in any field. But it is understandable that > having nothing between the disk access and something (lvm) should have a > performance penalty. > I know you can hack around nicely with disks and lvm, but those pro's > fall into the same category of questions people are suggesting related > to putting disks in raid. > > Let alone the risk that your are taking when there is going to be a > significant performance penalty: > https://www.researchgate.net/publication/284897601_LVM_in_the_Linux_environment_Performance_examination > https://hrcak.srce.hr/index.php?show=clanak&id_clanak_jezik=216661 > > > > -----Original Message----- > From: David Turner [mailto:drakonstein@xxxxxxxxx] > Sent: donderdag 31 mei 2018 23:48 > To: Marc Roos > Cc: ceph-users > Subject: Re: Why the change from ceph-disk to ceph-volume > and lvm? (and just not stick with direct disk access) > > Your question assumes that ceph-disk was a good piece of software. It > had a bug list a mile long and nobody working on it. A common example > was how simple it was to mess up any part of the dozens of components > that allowed an OSD to autostart on boot. One of the biggest problems > was when ceph-disk was doing it's thing and an OSD would take longer > than 3 minutes to start and ceph-disk would give up on it. > > That is a little bit about why a new solution was sought after and why > ceph-disk is being removed entirely. LVM was a choice made to implement > something other than partitions and udev magic while still incorporating > the information still needed from all of that in a better solution. > There has been a lot of talk about this on the ML. > > On Thu, May 31, 2018 at 5:23 PM Marc Roos <M.Roos@xxxxxxxxxxxxxxxxx> > wrote: > > > > What is the reasoning behind switching to lvm? Does it make sense > to go > through (yet) another layer to access the disk? Why creating this > dependency and added complexity? It is fine as it is, or not? > > > > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com