Re: scrub error with ceph

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Before issuing scrub, you may check if those scrub errors would point to one (or a small subset of) disk/OSD, and if so, did those objects put in a specified interval?

It is a large amount of scrub errors in a small cluster, which might be caused by some hardware issue ?

________________________________
> Date: Mon, 7 Dec 2015 14:15:07 -0700 
> From: erming@xxxxxxxxxxx 
> To: ceph-users@xxxxxxxxxxxxxx 
> Subject:  scrub error with ceph 
> 
> 
> Hi, 
> 
> I found there are 128 scrub errors in my ceph system. Checked with 
> health detail and found many pgs with stuck unclean issue. Should I 
> repair all of them? Or what I should do? 
> 
> [root@gcloudnet ~]# ceph -s 
> 
> cluster a4d0879f-abdc-4f9d-8a4b-53ce57d822f1 
> 
> health HEALTH_ERR 128 pgs inconsistent; 128 scrub errors; mds1: 
> Client HTRC:cephfs_data failing to respond to cache pressure; mds0: 
> Client physics-007:cephfs_data failing to respond to cache pressure; 
> pool 'cephfs_data' is full 
> 
> monmap e3: 3 mons at 
> {gcloudnet=xxx.xxx.xxx.xxx:6789/0,gcloudsrv1=xxx.xxx.xxx.xxx:6789/0,gcloudsrv2=xxx.xxx.xxx.xxx:6789/0}, 
> election epoch 178, quorum 0,1,2 gcloudnet,gcloudsrv1,gcloudsrv2 
> 
> mdsmap e51000: 2/2/2 up {0=gcloudsrv1=up:active,1=gcloudnet=up:active} 
> 
> osdmap e2821: 18 osds: 18 up, 18 in 
> 
> pgmap v10457877: 3648 pgs, 23 pools, 10501 GB data, 38688 kobjects 
> 
> 14097 GB used, 117 TB / 130 TB avail 
> 
> 6 active+clean+scrubbing+deep 
> 
> 3513 active+clean 
> 
> 128 active+clean+inconsistent 
> 
> 1 active+clean+scrubbing 
> 
> 
> P.S. I am increasing the pg and pgp numbers for cephfs_data pool. 
> 
> Thanks, 
> 
> Erming 
> 
> 
> 
> -- 
> 
> ---------------------------------------------------- 
> Erming Pei, Ph.D, Senior System Analyst 
> HPC Grid/Cloud Specialist, ComputeCanada/WestGrid 
> 
> Research Computing Group, IST 
> University of Alberta, Canada T6G 2H1 
> Email: Erming@xxxxxxxxxxx<mailto:Erming@xxxxxxxxxxx> Erming.Pei@xxxxxxx<mailto:Erming.Pei@xxxxxxx> 
> Tel. : +1 7804929914 Fax: +1 7804921729 
> 
> _______________________________________________ ceph-users mailing list 
> ceph-users@xxxxxxxxxxxxxx 
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
 		 	   		  
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux