Hi,
That seems very odd - what do the logs say for the osds with slow requests?
Thanks
On Tue, Nov 24, 2015 at 2:20 AM, Mika c <mika.leaf666@xxxxxxxxx> wrote:
Hi Sean,Yes, the cluster scrubbing status(scrub + deep scrub) is almost two weeks.And the result of execute `ceph pg dump | grep scrub` is empty.But command of "ceph health" show there is "16 pgs active+clean+scrubbing+deep, 2 pgs active+clean+scrubbing".I have 2 osds have slow requests warning.Is it releated?
Best wishes,
Mika
2015-11-23 17:59 GMT+08:00 Sean Redmond <sean.redmond1@xxxxxxxxx>:Hi Mika,Have the scubs been running for a long time? Can you see what pool they are running on? You can check using `ceph pg dump | grep scrub`ThanksOn Mon, Nov 23, 2015 at 9:32 AM, Mika c <mika.leaf666@xxxxxxxxx> wrote:_______________________________________________Hi cephers,We are facing a scrub issue. Our CEPH cluster is using Trusty / Hammer 0.94.1 and have almost 320 OSD disks on 10 nodes.And there are more than 30,000 PGs on cluster.The cluster works fine until last week. We found the cluster health status start display "active+clean+scrubbing+deep".Some PGs scrub ok but next second another PG start to scrubbing (or deep scrubbing) everyday.We did not change the parameters of scrub. It should scrub once per day and deep scrub once per week. Has anyone experiences with this incident?
Best wishes,
Mika
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com