Re: Fwd: Re: Blocked ops after change from filestore on HDD to bluestore on SDD

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Could you send your ceph.conf file over please? Are you setting any tunables for OSD or Bluestore currently?


From: ceph-users <ceph-users-bounces@xxxxxxxxxxxxxx> on behalf of Uwe Sauter <uwe.sauter.de@xxxxxxxxx>
Sent: Thursday, February 28, 2019 8:33 AM
To: Marc Roos; ceph-users; vitalif
Subject: Re: Fwd: Re: Blocked ops after change from filestore on HDD to bluestore on SDD
 
Do you have anything particular in mind? I'm using mdb backend with maxsize = 1GB but currently the files are only about 23MB.


>
> I am having quite a few openldap servers (slaves) running also, make
> sure to use proper caching that saves a lot of disk io. 
>
>
>
>
> -----Original Message-----
> Sent: 28 February 2019 13:56
> To: uwe.sauter.de@xxxxxxxxx; Uwe Sauter; Ceph Users
> Subject: *****SPAM***** Re: Fwd: Re: Blocked ops after
> change from filestore on HDD to bluestore on SDD
>
> "Advanced power loss protection" is in fact a performance feature, not a
> safety one.
>
>
> 28 февраля 2019 г. 13:03:51 GMT+03:00, Uwe Sauter
> <uwe.sauter.de@xxxxxxxxx> пишет:
>
>        Hi all,
>       
>        thanks for your insights.
>       
>        Eneko,
>       
>
>                We tried to use a Samsung 840 Pro SSD as OSD some time ago and
> it was a no-go; it wasn't that performance was bad, it
>                just didn't work for the kind of use of OSD. Any HDD was
> better than it (the disk was healthy and have been used in a
>                software raid-1 for a pair of years).
>               
>                I suggest you check first that your Samsung 860 Pro disks work
> well for Ceph. Also, how is your host's RAM?
>
>
>        As already mentioned the hosts each have 64GB RAM. Each host has 3
> SSDs for OSD usage. Each OSD is using about 1.3GB virtual
>        memory / 400MB residual memory.
>       
>       
>       
>        Joachim,
>       
>
>                I can only recommend the use of enterprise SSDs. We've tested
> many consumer SSDs in the past, including your SSDs. Many
>                of them are not suitable for long-term use and some weard out
> within 6 months.
>
>
>        Unfortunately I couldn't afford enterprise grade SSDs. But I
> suspect that my workload (about 20 VMs for our infrastructure, the
>        most IO demanding is probably LDAP) is light enough that wearout
> won't be a problem.
>       
>        The issue I'm seeing then is probably related to direct IO if using
> bluestore. But with filestore, the file system cache probably
>        hides the latency issues.
>       
>       
>        Igor,
>       
>
>                AFAIR Samsung 860 Pro isn't for enterprise market, you
> shouldn't use consumer SSDs for Ceph.
>               
>                I had some experience with Samsung 960 Pro a while ago and it
> turned out that it handled fsync-ed writes very slowly
>                (comparing to the original/advertised performance). Which one
> can probably explain by the lack of power loss protection
>                for these drives. I suppose it's the same in your case.
>               
>                Here are a couple links on the topic:
>               
>               
> https://www.percona.com/blog/2018/02/08/fsync-performance-storage-devices/
>               
>               
> https://www.sebastien-han.fr/blog/2014/10/10/ceph-how-to-test-if-your-ssd-is-suitable-as-a-journal-device/
>
>
>        Power loss protection wasn't a criteria for me as the cluster hosts
> are distributed in two buildings with separate battery backed
>        UPSs. As mentioned above I suspect the main difference for my case
> between filestore and bluestore is file system cache vs. direct
>        IO. Which means I will keep using filestore.
>       
>        Regards,
>       
>                Uwe
> ________________________________
>
>        ceph-users mailing list
>        ceph-users@xxxxxxxxxxxxxx
>        http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
> --
> With best regards,
> Vitaliy Filippov
>
>

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux