Scrub proces, IO performance

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello,

We are using Ceph as a storage backend for KVM, used for hosting MS Windows RDP, Linux for web applications with MySQL database and file sharing from Linux. Wen scrub or deep-scrub process is active, RDP sessions are freezing for a few seconds and web applications have big replay latency.

New we have disabled scrubbing  and deep-scrubbing process between  6AM - 10PM, when majority of users doesn't work, but user experience is still poor, like I write above. We are considering disabling scrubbing process at all. Does a new version 0.87 with addresses scrubbing priority is going to solve our problem (according to http://tracker.ceph.com/issues/6278)? Can we switch off scrubbing at all? How we can change our configuration to lower scrubbing performance impact? Does changing block size  can lower scrubbing impact or increase performance?

 

Our Ceph cluster configuration :

 

* we are using ~216 RBD disks for KVM VM's

* ~11TB used, 3.593TB data, replica count 3

* we have 5 mons, 32 OSD

* 3 pools/ 4096pgs (only one - RBD in use)

* 6 nodes (5osd+mon, 1 osd only) in two racks

* 1 SATA disk for system, 1 SSD disk for journal and 4 or 6 SATA disk for OSD

* 2 networks on 2 NIC 1Gbps (cluster + public)  on all nodes.

* 2x 10GBps links between racks

* without scrub max 45 iops

* when scrub running 120 - 180 iops

 

 

ceph.conf

 

mon initial members = ceph35, ceph30, ceph20, ceph15, ceph10

mon host = 10.20.8.35, 10.20.8.30, 10.20.8.20, 10.20.8.15, 10.20.8.10

 

public network = 10.20.8.0/22

cluster network = 10.20.4.0/22

 

filestore xattr use omap = true

filestore max sync interval = 15

 

osd journal size = 10240

osd pool default size = 3

osd pool default min size = 1

osd pool default pg num = 2048

osd pool default pgp num = 2048

osd crush chooseleaf type = 1

osd recovery max active = 1

osd recovery op priority = 1

osd max backfills = 1

 

auth cluster required = cephx

auth service required = cephx

auth client required = cephx

 

rbd default format = 2

 

Regards,

Mateusz

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux