Before issuing scrub, you may check if those scrub errors would point to one (or a small subset of) disk/OSD, and if so, did those objects put in a specified interval? It is a large amount of scrub errors in a small cluster, which might be caused by some hardware issue ? ________________________________ > Date: Mon, 7 Dec 2015 14:15:07 -0700 > From: erming@xxxxxxxxxxx > To: ceph-users@xxxxxxxxxxxxxx > Subject: scrub error with ceph > > > Hi, > > I found there are 128 scrub errors in my ceph system. Checked with > health detail and found many pgs with stuck unclean issue. Should I > repair all of them? Or what I should do? > > [root@gcloudnet ~]# ceph -s > > cluster a4d0879f-abdc-4f9d-8a4b-53ce57d822f1 > > health HEALTH_ERR 128 pgs inconsistent; 128 scrub errors; mds1: > Client HTRC:cephfs_data failing to respond to cache pressure; mds0: > Client physics-007:cephfs_data failing to respond to cache pressure; > pool 'cephfs_data' is full > > monmap e3: 3 mons at > {gcloudnet=xxx.xxx.xxx.xxx:6789/0,gcloudsrv1=xxx.xxx.xxx.xxx:6789/0,gcloudsrv2=xxx.xxx.xxx.xxx:6789/0}, > election epoch 178, quorum 0,1,2 gcloudnet,gcloudsrv1,gcloudsrv2 > > mdsmap e51000: 2/2/2 up {0=gcloudsrv1=up:active,1=gcloudnet=up:active} > > osdmap e2821: 18 osds: 18 up, 18 in > > pgmap v10457877: 3648 pgs, 23 pools, 10501 GB data, 38688 kobjects > > 14097 GB used, 117 TB / 130 TB avail > > 6 active+clean+scrubbing+deep > > 3513 active+clean > > 128 active+clean+inconsistent > > 1 active+clean+scrubbing > > > P.S. I am increasing the pg and pgp numbers for cephfs_data pool. > > Thanks, > > Erming > > > > -- > > ---------------------------------------------------- > Erming Pei, Ph.D, Senior System Analyst > HPC Grid/Cloud Specialist, ComputeCanada/WestGrid > > Research Computing Group, IST > University of Alberta, Canada T6G 2H1 > Email: Erming@xxxxxxxxxxx<mailto:Erming@xxxxxxxxxxx> Erming.Pei@xxxxxxx<mailto:Erming.Pei@xxxxxxx> > Tel. : +1 7804929914 Fax: +1 7804921729 > > _______________________________________________ ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com