Re: Inconsistent PGs every few days

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

I run a cluster with 7 OSD. The cluster has no much traffic on it. But 
every few days, I get a HEALTH_ERR, because of inconsistent PGs:

root at Sam ~ # ceph status
   cluster:
     id:     c4bfc288-8ba8-4c3a-b3a6-ed95503f50b7
     health: HEALTH_ERR
             3 scrub errors
             Possible data damage: 3 pgs inconsistent

   services:
     mon: 1 daemons, quorum mon1
     mgr: ceph-osd1(active)
     mds: FS-1/1/1 up  {0=ceph-osd1=up:active}
     osd: 11 osds: 8 up, 7 in
     rgw: 1 daemon active

   data:
     pools:   6 pools, 168 pgs
     objects: 901.8 k objects, 2.6 TiB
     usage: 7.9 TiB used, 7.4 TiB / 15 TiB avail
     pgs:     165 active+clean
              3   active+clean+inconsistent

   io:
     client:   641 KiB/s wr, 0 op/s rd, 3 op/s wr


root at Sam ~ # ceph health detail
HEALTH_ERR 3 scrub errors; Possible data damage: 3 pgs inconsistent
OSD_SCRUB_ERRORS 3 scrub errors
PG_DAMAGED Possible data damage: 3 pgs inconsistent
     pg 5.1d is active+clean+inconsistent, acting [6,8,3]
     pg 5.20 is active+clean+inconsistent, acting [3,9,0]
     pg 5.4a is active+clean+inconsistent, acting [6,3,7]

What's the reason for this problem? How can I analyse it?


Dimitri, what is your hardware? Your RAM is ECC?



k

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux