Re: Fwd: Re: Blocked ops after change from filestore on HDD to bluestore on SDD

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I already sent my configuration to the list about 3,5h ago but here it is again:


[global]
  auth client required = cephx
  auth cluster required = cephx
  auth service required = cephx
  cluster network = 169.254.42.0/24
  fsid = 753c9bbd-74bd-4fea-8c1e-88da775c5ad4
  keyring = /etc/pve/priv/$cluster.$name.keyring
  public network = 169.254.42.0/24

[mon]
  mon allow pool delete = true
  mon data avail crit = 5
  mon data avail warn = 15

[osd]
  keyring = /var/lib/ceph/osd/ceph-$id/keyring
  osd journal size = 5120
  osd pool default min size = 2
  osd pool default size = 3
  osd max backfills = 6
  osd recovery max active = 12

[mon.px-golf-cluster]
  host = px-golf-cluster
  mon addr = 169.254.42.54:6789

[mon.px-hotel-cluster]
  host = px-hotel-cluster
  mon addr = 169.254.42.55:6789

[mon.px-india-cluster]
  host = px-india-cluster
  mon addr = 169.254.42.56:6789



Am 28.02.19 um 14:44 schrieb Matthew H:
> Could you send your ceph.conf file over please? Are you setting any tunables for OSD or Bluestore currently?
> 
> ----------------------------------------------------------------------------------------------------------------------------------
> *From:* ceph-users <ceph-users-bounces@xxxxxxxxxxxxxx> on behalf of Uwe Sauter <uwe.sauter.de@xxxxxxxxx>
> *Sent:* Thursday, February 28, 2019 8:33 AM
> *To:* Marc Roos; ceph-users; vitalif
> *Subject:* Re:  Fwd: Re: Blocked ops after change from filestore on HDD to bluestore on SDD
>  
> Do you have anything particular in mind? I'm using mdb backend with maxsize = 1GB but currently the files are only about 23MB.
> 
> 
>> 
>> I am having quite a few openldap servers (slaves) running also, make 
>> sure to use proper caching that saves a lot of disk io.  
>> 
>> 
>> 
>> 
>> -----Original Message-----
>> Sent: 28 February 2019 13:56
>> To: uwe.sauter.de@xxxxxxxxx; Uwe Sauter; Ceph Users
>> Subject: *****SPAM***** Re:  Fwd: Re: Blocked ops after 
>> change from filestore on HDD to bluestore on SDD
>> 
>> "Advanced power loss protection" is in fact a performance feature, not a 
>> safety one.
>> 
>> 
>> 28 февраля 2019 г. 13:03:51 GMT+03:00, Uwe Sauter 
>> <uwe.sauter.de@xxxxxxxxx> пишет:
>> 
>>        Hi all,
>>        
>>        thanks for your insights.
>>        
>>        Eneko,
>>        
>> 
>>                We tried to use a Samsung 840 Pro SSD as OSD some time ago and 
>> it was a no-go; it wasn't that performance was bad, it 
>>                just didn't work for the kind of use of OSD. Any HDD was 
>> better than it (the disk was healthy and have been used in a 
>>                software raid-1 for a pair of years).
>>                
>>                I suggest you check first that your Samsung 860 Pro disks work 
>> well for Ceph. Also, how is your host's RAM?
>> 
>> 
>>        As already mentioned the hosts each have 64GB RAM. Each host has 3 
>> SSDs for OSD usage. Each OSD is using about 1.3GB virtual
>>        memory / 400MB residual memory.
>>        
>>        
>>        
>>        Joachim,
>>        
>> 
>>                I can only recommend the use of enterprise SSDs. We've tested 
>> many consumer SSDs in the past, including your SSDs. Many 
>>                of them are not suitable for long-term use and some weard out 
>> within 6 months.
>> 
>> 
>>        Unfortunately I couldn't afford enterprise grade SSDs. But I 
>> suspect that my workload (about 20 VMs for our infrastructure, the
>>        most IO demanding is probably LDAP) is light enough that wearout 
>> won't be a problem.
>>        
>>        The issue I'm seeing then is probably related to direct IO if using 
>> bluestore. But with filestore, the file system cache probably
>>        hides the latency issues.
>>        
>>        
>>        Igor,
>>        
>> 
>>                AFAIR Samsung 860 Pro isn't for enterprise market, you 
>> shouldn't use consumer SSDs for Ceph.
>>                
>>                I had some experience with Samsung 960 Pro a while ago and it 
>> turned out that it handled fsync-ed writes very slowly 
>>                (comparing to the original/advertised performance). Which one 
>> can probably explain by the lack of power loss protection 
>>                for these drives. I suppose it's the same in your case.
>>                
>>                Here are a couple links on the topic:
>>                
>>                
>> https://www.percona.com/blog/2018/02/08/fsync-performance-storage-devices/
>>                
>>                
>> https://www.sebastien-han.fr/blog/2014/10/10/ceph-how-to-test-if-your-ssd-is-suitable-as-a-journal-device/
>> 
>> 
>>        Power loss protection wasn't a criteria for me as the cluster hosts 
>> are distributed in two buildings with separate battery backed
>>        UPSs. As mentioned above I suspect the main difference for my case 
>> between filestore and bluestore is file system cache vs. direct
>>        IO. Which means I will keep using filestore.
>>        
>>        Regards,
>>        
>>                Uwe
>> ________________________________
>> 
>>        ceph-users mailing list
>>        ceph-users@xxxxxxxxxxxxxx
>>        http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> 
>> 
>> --
>> With best regards,
>> Vitaliy Filippov
>> 
>> 
> 
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux