We find this as well in our fresh built Jewel clusters, and seems to happen only with a handful of PGs from couple of pools. Thanks! On 9/21/16, 3:14 PM, "ceph-users on behalf of Tobias Böhm" <ceph-users-bounces@xxxxxxxxxxxxxx on behalf of tb@xxxxxxxxxx> wrote: Hi, there is an open bug in the tracker: http://tracker.ceph.com/issues/16474 It also suggests restarting OSDs as a workaround. We faced the same issue after increasing the number of PGs in our cluster and restarting OSDs solved it as well. Tobias > Am 21.09.2016 um 11:26 schrieb Dan van der Ster <dan@xxxxxxxxxxxxxx>: > > There was a thread about this a few days ago: > http://lists.ceph.com/pipermail/ceph-users-ceph.com/2016-September/012857.html > And the OP found a workaround. > Looks like a bug though... (by default PGs scrub at most once per day). > > -- dan > > > > On Tue, Sep 20, 2016 at 10:43 PM, Martin Bureau <mbureau@xxxxxxxxxxxx> wrote: >> Hello, >> >> >> I noticed that the same pg gets scrubbed repeatedly on our new Jewel >> cluster: >> >> >> Here's an excerpt from log: >> >> >> 2016-09-20 20:36:31.236123 osd.12 10.1.82.82:6820/14316 150514 : cluster >> [INF] 25.3f scrub ok >> 2016-09-20 20:36:32.232918 osd.12 10.1.82.82:6820/14316 150515 : cluster >> [INF] 25.3f scrub starts >> 2016-09-20 20:36:32.236876 osd.12 10.1.82.82:6820/14316 150516 : cluster >> [INF] 25.3f scrub ok >> 2016-09-20 20:36:33.233268 osd.12 10.1.82.82:6820/14316 150517 : cluster >> [INF] 25.3f deep-scrub starts >> 2016-09-20 20:36:33.242258 osd.12 10.1.82.82:6820/14316 150518 : cluster >> [INF] 25.3f deep-scrub ok >> 2016-09-20 20:36:36.233604 osd.12 10.1.82.82:6820/14316 150519 : cluster >> [INF] 25.3f scrub starts >> 2016-09-20 20:36:36.237221 osd.12 10.1.82.82:6820/14316 150520 : cluster >> [INF] 25.3f scrub ok >> 2016-09-20 20:36:41.234490 osd.12 10.1.82.82:6820/14316 150521 : cluster >> [INF] 25.3f deep-scrub starts >> 2016-09-20 20:36:41.243720 osd.12 10.1.82.82:6820/14316 150522 : cluster >> [INF] 25.3f deep-scrub ok >> 2016-09-20 20:36:45.235128 osd.12 10.1.82.82:6820/14316 150523 : cluster >> [INF] 25.3f deep-scrub starts >> 2016-09-20 20:36:45.352589 osd.12 10.1.82.82:6820/14316 150524 : cluster >> [INF] 25.3f deep-scrub ok >> 2016-09-20 20:36:47.235310 osd.12 10.1.82.82:6820/14316 150525 : cluster >> [INF] 25.3f scrub starts >> 2016-09-20 20:36:47.239348 osd.12 10.1.82.82:6820/14316 150526 : cluster >> [INF] 25.3f scrub ok >> 2016-09-20 20:36:49.235538 osd.12 10.1.82.82:6820/14316 150527 : cluster >> [INF] 25.3f deep-scrub starts >> 2016-09-20 20:36:49.243121 osd.12 10.1.82.82:6820/14316 150528 : cluster >> [INF] 25.3f deep-scrub ok >> 2016-09-20 20:36:51.235956 osd.12 10.1.82.82:6820/14316 150529 : cluster >> [INF] 25.3f deep-scrub starts >> 2016-09-20 20:36:51.244201 osd.12 10.1.82.82:6820/14316 150530 : cluster >> [INF] 25.3f deep-scrub ok >> 2016-09-20 20:36:52.236076 osd.12 10.1.82.82:6820/14316 150531 : cluster >> [INF] 25.3f scrub starts >> 2016-09-20 20:36:52.239376 osd.12 10.1.82.82:6820/14316 150532 : cluster >> [INF] 25.3f scrub ok >> 2016-09-20 20:36:56.236740 osd.12 10.1.82.82:6820/14316 150533 : cluster >> [INF] 25.3f scrub starts >> >> >> How can I troubleshoot / resolve this ? >> >> >> Regards, >> >> Martin >> >> >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@xxxxxxxxxxxxxx >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com