Furthermore, presuming you are running Jewel or Luminous you can change some settings in ceph.conf to mitigate the deep-scrub impact:
osd scrub max interval = 4838400
osd scrub min interval = 2419200
osd scrub interval randomize ratio = 1.0
osd scrub chunk max = 1
osd scrub chunk min = 1
osd scrub priority = 1
osd scrub sleep = 0.1
osd deep scrub interval = 2419200
osd deep scrub stride = 1048576
osd disk thread ioprio class = idle
osd disk thread ioprio priority = 7
osd scrub max interval = 4838400
osd scrub min interval = 2419200
osd scrub interval randomize ratio = 1.0
osd scrub chunk max = 1
osd scrub chunk min = 1
osd scrub priority = 1
osd scrub sleep = 0.1
osd deep scrub interval = 2419200
osd deep scrub stride = 1048576
osd disk thread ioprio class = idle
osd disk thread ioprio priority = 7
Kind regards,
Caspar
Caspar
Op ma 10 dec. 2018 om 12:06 schreef Vladimir Prokofev <v@xxxxxxxxxxx>:
_______________________________________________Hello list.Deep scrub totally kills cluster performance.First of all, it takes several minutes to complete:2018-12-09 01:39:53.857994 7f2d32fde700 0 log_channel(cluster) log [DBG] : 4.75 deep-scrub starts2018-12-09 01:46:30.703473 7f2d32fde700 0 log_channel(cluster) log [DBG] : 4.75 deep-scrub okSecond, while it runs, it consumes 100% of OSD time[1]. This is on an ordinary 7200RPM spinner.While this happens, VMs cannot access their disks, and that leads to service interruptions.I disabled scrub and deep-scrub operations for now, and have 2 major questions:- can I disable 'health warning' status for noscrub and nodeep-scrub? I thought there was a way to do this, but can't find it. I want my cluster to think it's healthy, so if any new 'slow requests' or anything else pops - it will change status to 'health warning' again;- is there a way to limit deepscrub impact on disk performance, or do I just have to go and buy SSDs?
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com