olcDbShmKey only applies to BDB and HDB backends but I'm using the new MDB backend. Am 28.02.19 um 14:47 schrieb Marc Roos: > If you have every second disk io with your current settings, which I > also had with 'default' settings. There are some optimizations you can > do, bringing it down to every 50 seconds or so. Adding the olcDbShmKey > will allow for slapd to access the db cache. > I am getting an error of sharedmemory settings when rebooting (centos7), > but maintainers of slapd said that I can ignore that. Dont have any > problems since using this also. > > > > -----Original Message----- > From: Uwe Sauter [mailto:uwe.sauter.de@xxxxxxxxx] > Sent: 28 February 2019 14:34 > To: Marc Roos; ceph-users; vitalif > Subject: Re: Fwd: Re: Blocked ops after change from > filestore on HDD to bluestore on SDD > > Do you have anything particular in mind? I'm using mdb backend with > maxsize = 1GB but currently the files are only about 23MB. > > >> >> I am having quite a few openldap servers (slaves) running also, make >> sure to use proper caching that saves a lot of disk io. >> >> >> >> >> -----Original Message----- >> Sent: 28 February 2019 13:56 >> To: uwe.sauter.de@xxxxxxxxx; Uwe Sauter; Ceph Users >> Subject: *****SPAM***** Re: Fwd: Re: Blocked ops after >> change from filestore on HDD to bluestore on SDD >> >> "Advanced power loss protection" is in fact a performance feature, not > >> a safety one. >> >> >> 28 февраля 2019 г. 13:03:51 GMT+03:00, Uwe Sauter >> <uwe.sauter.de@xxxxxxxxx> пишет: >> >> Hi all, >> >> thanks for your insights. >> >> Eneko, >> >> >> We tried to use a Samsung 840 Pro SSD as OSD some time ago > and it >> was a no-go; it wasn't that performance was bad, it >> just didn't work for the kind of use of OSD. Any HDD was > better than >> it (the disk was healthy and have been used in a >> software raid-1 for a pair of years). >> >> I suggest you check first that your Samsung 860 Pro disks > work well >> for Ceph. Also, how is your host's RAM? >> >> >> As already mentioned the hosts each have 64GB RAM. Each host has > 3 >> SSDs for OSD usage. Each OSD is using about 1.3GB virtual >> memory / 400MB residual memory. >> >> >> >> Joachim, >> >> >> I can only recommend the use of enterprise SSDs. We've > tested many >> consumer SSDs in the past, including your SSDs. Many >> of them are not suitable for long-term use and some weard > out within >> 6 months. >> >> >> Unfortunately I couldn't afford enterprise grade SSDs. But I > suspect >> that my workload (about 20 VMs for our infrastructure, the >> most IO demanding is probably LDAP) is light enough that wearout >> won't be a problem. >> >> The issue I'm seeing then is probably related to direct IO if > using >> bluestore. But with filestore, the file system cache probably >> hides the latency issues. >> >> >> Igor, >> >> >> AFAIR Samsung 860 Pro isn't for enterprise market, you > shouldn't use >> consumer SSDs for Ceph. >> >> I had some experience with Samsung 960 Pro a while ago and > it turned >> out that it handled fsync-ed writes very slowly >> (comparing to the original/advertised performance). Which > one can >> probably explain by the lack of power loss protection >> for these drives. I suppose it's the same in your case. >> >> Here are a couple links on the topic: >> >> >> https://www.percona.com/blog/2018/02/08/fsync-performance-storage-devi >> ces/ >> >> >> https://www.sebastien-han.fr/blog/2014/10/10/ceph-how-to-test-if-your- >> ssd-is-suitable-as-a-journal-device/ >> >> >> Power loss protection wasn't a criteria for me as the cluster > hosts >> are distributed in two buildings with separate battery backed >> UPSs. As mentioned above I suspect the main difference for my > case >> between filestore and bluestore is file system cache vs. direct >> IO. Which means I will keep using filestore. >> >> Regards, >> >> Uwe >> ________________________________ >> >> ceph-users mailing list >> ceph-users@xxxxxxxxxxxxxx >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> >> -- >> With best regards, >> Vitaliy Filippov >> >> > > > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com